You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@apex.apache.org by vr...@apache.org on 2017/12/28 17:54:01 UTC
[apex-malhar] 02/02: APEXMALHAR-2517 imports/checkstyle #662
This is an automated email from the ASF dual-hosted git repository.
vrozov pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/apex-malhar.git
commit 0d98d05ef670dcf8359681b384de1d78734f6657
Author: Apex Dev <de...@apex.apache.org>
AuthorDate: Mon Aug 14 06:53:10 2017 -0700
APEXMALHAR-2517 imports/checkstyle #662
---
apps/filecopy/pom.xml | 4 +-
.../com/datatorrent/apps/copy/HDFSFileCopyApp.java | 4 +-
apps/logstream/pom.xml | 2 +-
.../logstream/AggregationsToRedisOperator.java | 4 +-
.../datatorrent/apps/logstream/Application.java | 29 +--
.../datatorrent/apps/logstream/Application1.java | 5 +-
.../apps/logstream/DimensionOperator.java | 5 +-
.../apps/logstream/DimensionOperatorUnifier.java | 3 +-
.../apps/logstream/LogScoreOperator.java | 4 +-
.../datatorrent/apps/logstream/LogstreamTopN.java | 6 +-
.../logstream/LogstreamWidgetOutputOperator.java | 5 +-
.../apps/logstream/RabbitMQLogsInputOperator.java | 3 +-
.../apps/logstream/DimensionOperatorTest.java | 5 +-
.../logstream/DimensionOperatorUnifierTest.java | 5 +-
.../apps/logstream/FilterOperatorTest.java | 2 +-
.../apps/logstream/LogstreamTopNTest.java | 5 +-
apps/pom.xml | 2 +-
benchmark/pom.xml | 2 +-
.../apache/apex/benchmark/ApplicationFixed.java | 2 +-
.../java/org/apache/apex/benchmark/Benchmark.java | 2 +-
.../apache/apex/benchmark/CouchBaseAppInput.java | 2 +-
.../apache/apex/benchmark/CouchBaseAppOutput.java | 4 +-
.../apex/benchmark/CouchBaseInputOperator.java | 6 +-
.../apex/benchmark/CouchBaseOutputOperator.java | 4 +-
.../apex/benchmark/FixedTuplesInputOperator.java | 2 +-
.../org/apache/apex/benchmark/RandomMapOutput.java | 2 +-
.../apex/benchmark/RandomWordInputModule.java | 2 +-
.../apache/apex/benchmark/WordCountOperator.java | 2 +-
.../AerospikeOutputBenchmarkApplication.java | 6 +-
.../aerospike/AerospikeOutputOperator.java | 6 +-
.../algo/UniqueValueCountBenchmarkApplication.java | 14 +-
.../CassandraOutputBenchmarkApplication.java | 8 +-
.../cassandra/CassandraOutputOperator.java | 6 +-
.../apex/benchmark/fs/FSByteOutputOperator.java | 4 +-
.../benchmark/fs/FSOutputOperatorBenchmark.java | 8 +-
.../benchmark/hive/HiveInsertBenchmarkingApp.java | 8 +-
.../hive/HiveMapInsertBenchmarkingApp.java | 13 +-
.../kafka/BenchmarkKafkaInputOperator.java | 5 +-
.../BenchmarkPartitionableKafkaOutputOperator.java | 2 +-
.../apex/benchmark/kafka/KafkaInputBenchmark.java | 9 +-
.../apex/benchmark/kafka/KafkaOutputBenchmark.java | 2 +-
.../apex/benchmark/kafka/KafkaTestPartitioner.java | 2 +-
.../org/apache/apex/benchmark/package-info.java | 2 +-
.../script/RubyOperatorBenchmarkApplication.java | 12 +-
.../benchmark/spillable/SpillableBenchmarkApp.java | 4 +-
.../spillable/SpillableTestInputOperator.java | 2 +-
.../benchmark/spillable/SpillableTestOperator.java | 2 +-
.../benchmark/state/ManagedStateBenchmarkApp.java | 6 +-
.../apache/apex/benchmark/state/StoreOperator.java | 4 +-
.../benchmark/stream/DevNullCounterBenchmark.java | 6 +-
.../apex/benchmark/stream/IntegerOperator.java | 2 +-
.../apex/benchmark/stream/StreamDuplicaterApp.java | 6 +-
.../apex/benchmark/stream/StreamMergeApp.java | 6 +-
.../benchmark/testbench/EventClassifierApp.java | 6 +-
.../EventClassifierNumberToHashDoubleApp.java | 8 +-
.../benchmark/testbench/EventGeneratorApp.java | 6 +-
.../benchmark/testbench/EventIncrementerApp.java | 6 +-
.../benchmark/testbench/FilterClassifierApp.java | 6 +-
.../testbench/FilteredEventClassifierApp.java | 6 +-
.../apex/benchmark/testbench/HashMapOperator.java | 5 +-
.../testbench/RandomEventGeneratorApp.java | 6 +-
.../benchmark/testbench/SeedEventGeneratorApp.java | 9 +-
.../benchmark/testbench/ThroughputCounterApp.java | 6 +-
.../apex/benchmark/window/AbstractGenerator.java | 2 +-
.../AbstractWindowedOperatorBenchmarkApp.java | 8 +-
.../window/KeyedWindowedOperatorBenchmarkApp.java | 6 +-
.../window/WindowedOperatorBenchmarkApp.java | 2 +-
.../apex/benchmark/ApplicationFixedTest.java | 2 +-
.../org/apache/apex/benchmark/BenchmarkTest.java | 2 +-
.../apex/benchmark/CouchBaseBenchmarkTest.java | 2 +-
.../apex/benchmark/accumulo/AccumuloApp.java | 10 +-
.../apex/benchmark/accumulo/AccumuloAppTest.java | 2 +-
.../aerospike/AerospikeBenchmarkAppTest.java | 2 +-
.../algo/UniqueValueCountBenchmarkTest.java | 2 +-
.../cassandra/CassandraApplicatonTest.java | 2 +-
.../apex/benchmark/hbase/HBaseApplicationTest.java | 2 +-
.../hbase/HBaseCsvMappingApplication.java | 6 +-
.../benchmark/hive/HiveInsertBenchmarkTest.java | 2 +-
.../apex/benchmark/hive/HiveMapBenchmarkTest.java | 2 +-
.../benchmark/kafka/KafkaInputBenchmarkTest.java | 2 +-
.../benchmark/kafka/KafkaOutputBenchmarkTest.java | 2 +-
.../benchmark/memsql/MemsqlInputBenchmark.java | 6 +-
.../benchmark/memsql/MemsqlInputBenchmarkTest.java | 16 +-
.../benchmark/memsql/MemsqlOutputBenchmark.java | 6 +-
.../memsql/MemsqlOutputBenchmarkTest.java | 6 +-
.../script/RubyOperatorBenchmarkAppTest.java | 2 +-
.../spillable/SpillableBenchmarkAppTester.java | 2 +-
.../spillable/SpillableDSBenchmarkTest.java | 5 +-
.../state/ManagedStateBenchmarkAppTest.java | 4 +-
.../testbench/EventClassifierAppTest.java | 2 +-
.../EventClassifierNumberToHashDoubleAppTest.java | 2 +-
.../benchmark/testbench/EventGeneratorAppTest.java | 2 +-
.../testbench/EventIncrementerAppTest.java | 2 +-
.../testbench/FilterClassifierAppTest.java | 2 +-
.../testbench/FilteredEventClassifierAppTest.java | 2 +-
.../testbench/ThroughputCounterAppTest.java | 2 +-
.../util/serde/GenericSerdePerformanceTest.java | 2 +-
.../KeyedWindowedOperatorBenchmarkAppTest.java | 2 +-
.../window/WindowedOperatorBenchmarkAppTest.java | 2 +-
contrib/pom.xml | 11 +-
.../accumulo/AbstractAccumuloInputOperator.java | 13 +-
.../accumulo/AbstractAccumuloOutputOperator.java | 26 +-
.../malhar/contrib/accumulo/AccumuloStore.java | 70 ++++--
.../contrib/accumulo/AccumuloWindowStore.java | 71 +++---
.../apex/malhar/contrib/accumulo/package-info.java | 2 +-
.../aerospike/AbstractAerospikeGetOperator.java | 32 +--
...stractAerospikeNonTransactionalPutOperator.java | 30 +--
.../AbstractAerospikeTransactionalPutOperator.java | 33 +--
.../AerospikePOJONonTransactionalPutOperator.java | 20 +-
.../AerospikePOJOTransactionalPutOperator.java | 24 +-
.../malhar/contrib/aerospike/AerospikeStore.java | 48 ++--
.../aerospike/AerospikeTransactionalStore.java | 87 ++++---
.../malhar/contrib/aerospike/package-info.java | 2 +-
.../contrib/apachelog/ApacheLogInputGenerator.java | 44 ++--
.../contrib/apachelog/ApplicationLocalLog.java | 11 +-
.../contrib/apachelog/ApplicationLogGenerator.java | 14 +-
.../malhar/contrib/apachelog/GeoIPExtractor.java | 18 +-
.../contrib/apachelog/TimestampExtractor.java | 11 +-
.../contrib/apachelog/UserAgentExtractor.java | 10 +-
.../malhar/contrib/apachelog/package-info.java | 2 +-
.../malhar/contrib/apachelog/zmq/Application.java | 16 +-
.../malhar/contrib/avro/AvroFileInputOperator.java | 4 +-
.../malhar}/contrib/avro/AvroFileToPojoModule.java | 2 +-
.../apex/malhar/contrib/avro/AvroRecordHelper.java | 2 +-
.../apex/malhar/contrib/avro/AvroToPojo.java | 10 +-
.../apex/malhar/contrib/avro/PojoToAvro.java | 9 +-
.../cassandra/AbstractCassandraInputOperator.java | 14 +-
...ractCassandraTransactionableOutputOperator.java | 8 +-
.../cassandra/AbstractUpsertOutputOperator.java | 4 +-
.../cassandra/CassandraPOJOInputOperator.java | 25 +-
.../cassandra/CassandraPOJOOutputOperator.java | 47 ++--
.../contrib/cassandra/CassandraPojoUtils.java | 5 +-
.../CassandraPreparedStatementGenerator.java | 2 +-
.../malhar/contrib/cassandra/CassandraStore.java | 39 +--
.../cassandra/CassandraTransactionalStore.java | 45 ++--
.../contrib/cassandra/ConnectionStateManager.java | 4 +-
.../contrib/cassandra/UpsertExecutionContext.java | 4 +-
.../malhar/contrib/cassandra/package-info.java | 2 +-
.../couchbase/AbstractCouchBaseInputOperator.java | 34 ++-
.../couchbase/AbstractCouchBaseOutputOperator.java | 29 ++-
.../couchbase/AbstractCouchBaseSetOperator.java | 2 +-
.../contrib/couchbase/CouchBaseJSONSerializer.java | 7 +-
.../couchbase/CouchBasePOJOInputOperator.java | 20 +-
.../contrib/couchbase/CouchBaseSerializer.java | 2 +-
.../malhar/contrib/couchbase/CouchBaseStore.java | 34 ++-
.../contrib/couchbase/CouchBaseWindowStore.java | 33 ++-
.../couchbase/CouchbasePOJOSetOperator.java | 8 +-
.../malhar/contrib/couchbase/package-info.java | 2 +-
.../couchdb/AbstractCouchDBInputOperator.java | 10 +-
.../couchdb/AbstractCouchDBOutputOperator.java | 5 +-
.../couchdb/AbstractMapBasedInputOperator.java | 8 +-
.../contrib/couchdb/CouchDBPOJOInputOperator.java | 45 ++--
.../contrib/couchdb/CouchDBPOJOOutputOperator.java | 7 +-
.../apex/malhar/contrib/couchdb/CouchDbStore.java | 10 +-
.../couchdb/MapBasedCouchDbOutputOperator.java | 4 +-
.../apex/malhar/contrib/couchdb/package-info.java | 2 +-
.../AbstractElasticSearchInputOperator.java | 7 +-
.../AbstractElasticSearchOutputOperator.java | 30 +--
.../elasticsearch/ElasticSearchConnectable.java | 10 +-
.../ElasticSearchMapInputOperator.java | 8 +-
.../ElasticSearchMapOutputOperator.java | 10 +-
.../ElasticSearchPercolatorOperator.java | 10 +-
.../ElasticSearchPercolatorStore.java | 5 +-
.../malhar/contrib/elasticsearch/package-info.java | 2 +-
.../malhar/contrib/enrich/AbstractEnricher.java | 12 +-
.../apex/malhar/contrib/enrich/BackendLoader.java | 7 +-
.../malhar/contrib/enrich/DelimitedFSLoader.java | 11 +-
.../apex/malhar/contrib/enrich/FSLoader.java | 6 +-
.../malhar/contrib/enrich/FixedWidthFSLoader.java | 7 +-
.../apex/malhar/contrib/enrich/JDBCLoader.java | 6 +-
.../apex/malhar/contrib/enrich/JsonFSLoader.java | 2 +-
.../apex/malhar/contrib/enrich/MapEnricher.java | 6 +-
.../contrib/enrich/NullValuesCacheManager.java | 5 +-
.../apex/malhar/contrib/enrich/POJOEnricher.java | 13 +-
.../apex/malhar/contrib/enrich/ReadOnlyBackup.java | 5 +-
.../malhar/contrib/formatter/CsvFormatter.java | 11 +-
.../contrib/geode/AbstractGeodeInputOperator.java | 4 +-
.../contrib/geode/AbstractGeodeOutputOperator.java | 4 +-
.../malhar/contrib/geode/GeodeCheckpointStore.java | 27 ++-
.../contrib/geode/GeodeKeyValueStorageAgent.java | 5 +-
.../contrib/geode/GeodePOJOOutputOperator.java | 11 +-
.../apex/malhar/contrib/geode/GeodeStore.java | 10 +-
.../malhar/contrib/geode/RegionCreateFunction.java | 2 +-
.../hbase/AbstractHBaseAppendOutputOperator.java | 13 +-
.../contrib/hbase/AbstractHBaseOutputOperator.java | 14 +-
.../hbase/AbstractHBasePutOutputOperator.java | 8 +-
.../AbstractHBaseWindowAppendOutputOperator.java | 28 ++-
.../hbase/AbstractHBaseWindowOutputOperator.java | 6 +-
.../AbstractHBaseWindowPutOutputOperator.java | 31 +--
.../contrib/hbase/HBaseCsvMappingPutOperator.java | 44 ++--
.../apex/malhar/contrib/hbase/HBaseFieldInfo.java | 133 +++++-----
.../contrib/hbase/HBaseFieldValueGenerator.java | 6 +-
.../malhar/contrib/hbase/HBaseGetOperator.java | 2 +-
.../malhar/contrib/hbase/HBaseInputOperator.java | 5 +-
.../hbase/HBaseNameValueCsvPutOperator.java | 71 +++---
.../malhar/contrib/hbase/HBaseOperatorBase.java | 11 +-
.../contrib/hbase/HBasePOJOInputOperator.java | 28 +--
.../malhar/contrib/hbase/HBasePOJOPutOperator.java | 13 +-
.../malhar/contrib/hbase/HBaseScanOperator.java | 32 +--
.../apex/malhar/contrib/hbase/HBaseStore.java | 80 +++---
.../malhar/contrib/hbase/HBaseWindowStore.java | 55 +++--
.../apex/malhar}/contrib/hbase/OutputAdapter.java | 4 +-
.../apex/malhar/contrib/hbase/package-info.java | 2 +-
.../apex/malhar/contrib/jython/PythonOperator.java | 18 +-
.../apex/malhar/contrib/jython/package-info.java | 2 +-
.../AbstractExactlyOnceKafkaOutputOperator.java | 36 ++-
.../contrib/kafka/AbstractKafkaInputOperator.java | 269 ++++++++++-----------
.../contrib/kafka/AbstractKafkaOutputOperator.java | 18 +-
.../AbstractKafkaSinglePortInputOperator.java | 2 +-
.../contrib/kafka/HighlevelKafkaConsumer.java | 8 +-
.../apex/malhar/contrib/kafka/KafkaConsumer.java | 116 ++++-----
.../malhar/contrib/kafka/KafkaJsonEncoder.java | 2 +-
.../malhar/contrib/kafka/KafkaMetadataUtil.java | 22 +-
.../apex/malhar/contrib/kafka/KafkaPartition.java | 32 ++-
.../KafkaSinglePortByteArrayInputOperator.java | 32 ++-
.../kafka/KafkaSinglePortOutputOperator.java | 2 +-
.../kafka/KafkaSinglePortStringInputOperator.java | 8 +-
.../apex/malhar/contrib/kafka/OffsetManager.java | 2 +-
.../contrib/kafka/POJOKafkaOutputOperator.java | 8 +-
.../malhar/contrib/kafka/SimpleKafkaConsumer.java | 114 ++++-----
.../apex/malhar/contrib/kafka/package-info.java | 2 +-
.../kinesis/AbstractKinesisInputOperator.java | 194 +++++++--------
.../kinesis/AbstractKinesisOutputOperator.java | 43 ++--
.../kinesis/KinesisByteArrayInputOperator.java | 5 +-
.../kinesis/KinesisByteArrayOutputOperator.java | 2 +-
.../malhar/contrib/kinesis/KinesisConsumer.java | 62 +++--
.../kinesis/KinesisStringInputOperator.java | 8 +-
.../kinesis/KinesisStringOutputOperator.java | 3 +-
.../apex/malhar/contrib/kinesis/KinesisUtil.java | 41 ++--
.../apex/malhar/contrib/kinesis/ShardManager.java | 2 +-
.../apex/malhar/contrib/kinesis/package-info.java | 2 +-
.../memcache/AbstractMemcacheInputOperator.java | 4 +-
.../memcache/AbstractMemcacheOutputOperator.java | 7 +-
.../memcache/MemcachePOJOOutputOperator.java | 18 +-
.../malhar/contrib/memcache/MemcacheStore.java | 16 +-
.../apex/malhar/contrib/memcache/package-info.java | 2 +-
.../AbstractMemcacheInputOperator.java | 4 +-
.../AbstractMemcacheOutputOperator.java | 7 +-
.../contrib/memcache_whalin/MemcacheStore.java | 14 +-
.../contrib/memcache_whalin/package-info.java | 2 +-
.../memsql/AbstractMemsqlInputOperator.java | 4 +-
.../memsql/AbstractMemsqlOutputOperator.java | 4 +-
.../contrib/memsql/MemsqlPOJOInputOperator.java | 98 +++-----
.../contrib/memsql/MemsqlPOJOOutputOperator.java | 57 ++---
.../apex/malhar/contrib/memsql/MemsqlStore.java | 4 +-
.../apex/malhar/contrib/memsql/package-info.java | 2 +-
.../misc/algo/AbstractStreamPatternMatcher.java | 4 +-
.../malhar/contrib/misc/algo/AllAfterMatchMap.java | 4 +-
.../apex/malhar/contrib/misc/algo/DistinctMap.java | 6 +-
.../malhar/contrib/misc/algo/FilterKeyVals.java | 4 +-
.../contrib/misc/algo/FilterKeysHashMap.java | 4 +-
.../malhar/contrib/misc/algo/FilterKeysMap.java | 6 +-
.../malhar/contrib/misc/algo/FirstMatchMap.java | 4 +-
.../apex/malhar/contrib/misc/algo/FirstN.java | 3 +-
.../malhar/contrib/misc/algo/FirstTillMatch.java | 4 +-
.../malhar/contrib/misc/algo/InsertSortDesc.java | 5 +-
.../apex/malhar/contrib/misc/algo/InvertIndex.java | 4 +-
.../malhar/contrib/misc/algo/InvertIndexArray.java | 4 +-
.../malhar/contrib/misc/algo/LastMatchMap.java | 4 +-
.../contrib/misc/algo/LeastFrequentKeyMap.java | 8 +-
.../misc/algo/LeastFrequentKeyValueMap.java | 4 +-
.../contrib/misc/algo/MostFrequentKeyMap.java | 8 +-
.../contrib/misc/algo/MostFrequentKeyValueMap.java | 4 +-
.../apex/malhar/contrib/misc/algo/Sampler.java | 4 +-
.../{aerospike => misc/algo}/package-info.java | 2 +-
.../apex/malhar/contrib/misc/math/Change.java | 3 +-
.../apex/malhar/contrib/misc/math/ChangeAlert.java | 5 +-
.../contrib/misc/math/ChangeAlertKeyVal.java | 4 +-
.../malhar/contrib/misc/math/ChangeAlertMap.java | 2 +-
.../malhar/contrib/misc/math/ChangeKeyVal.java | 5 +-
.../malhar/contrib/misc/math/CompareExceptMap.java | 5 +-
.../apex/malhar/contrib/misc/math/CompareMap.java | 3 +-
.../apex/malhar/contrib/misc/math/CountKeyVal.java | 6 +-
.../apex/malhar/contrib/misc/math/ExceptMap.java | 5 +-
.../apex/malhar/contrib/misc/math/Quotient.java | 3 +-
.../apex/malhar/contrib/misc/math/QuotientMap.java | 2 +-
.../apex/malhar/contrib/misc/math/SumCountMap.java | 6 +-
.../{aerospike => misc/math}/package-info.java | 2 +-
.../contrib/misc/streamquery/DeleteOperator.java | 3 +-
.../misc/streamquery/GroupByHavingOperator.java | 7 +-
.../misc/streamquery/InnerJoinOperator.java | 5 +-
.../contrib/misc/streamquery/SelectOperator.java | 5 +-
.../contrib/misc/streamquery/UpdateOperator.java | 3 +-
.../streamquery/condition/BetweenCondition.java | 2 +-
.../streamquery/condition/CompoundCondition.java | 2 +-
.../streamquery/condition/EqualValueCondition.java | 2 +-
.../misc/streamquery/condition/InCondition.java | 2 +-
.../misc/streamquery/condition/LikeCondition.java | 2 +-
.../misc/streamquery/index/BinaryExpression.java | 2 +-
.../contrib/misc/streamquery/index/MidIndex.java | 2 +-
.../misc/streamquery/index/NegateExpression.java | 2 +-
.../misc/streamquery/index/RoundDoubleIndex.java | 2 +-
.../misc/streamquery/index/StringCaseIndex.java | 2 +-
.../misc/streamquery/index/StringLenIndex.java | 2 +-
.../misc/streamquery/index/SumExpression.java | 2 +-
.../misc/streamquery/index/UnaryExpression.java | 2 +-
.../mongodb/MongoDBArrayListOutputOperator.java | 7 +-
.../malhar/contrib/mongodb/MongoDBConnectable.java | 13 +-
.../mongodb/MongoDBHashMapOutputOperator.java | 9 +-
.../contrib/mongodb/MongoDBInputOperator.java | 21 +-
.../contrib/mongodb/MongoDBOutputOperator.java | 56 ++---
.../contrib/mongodb/MongoDBPOJOOutputOperator.java | 27 +--
.../apex/malhar/contrib/mongodb/package-info.java | 2 +-
.../contrib/mqtt/AbstractMqttInputOperator.java | 30 +--
.../contrib/mqtt/AbstractMqttOutputOperator.java | 13 +-
.../mqtt/AbstractSinglePortMqttInputOperator.java | 6 +-
.../mqtt/AbstractSinglePortMqttOutputOperator.java | 2 +-
.../apex/malhar/contrib/mqtt/MqttClientConfig.java | 4 +-
.../apex/malhar/contrib/mqtt/package-info.java | 2 +-
.../contrib/nifi/AbstractNiFiInputOperator.java | 5 +-
.../contrib/nifi/AbstractNiFiOutputOperator.java | 2 +-
.../nifi/AbstractNiFiSinglePortInputOperator.java | 2 +-
.../apex/malhar/contrib/nifi/NiFiDataPacket.java | 2 +-
.../malhar/contrib/nifi/NiFiDataPacketBuilder.java | 2 +-
.../contrib/nifi/NiFiSinglePortInputOperator.java | 2 +-
.../contrib/nifi/NiFiSinglePortOutputOperator.java | 2 +-
.../contrib/nifi/StandardNiFiDataPacket.java | 2 +-
.../contrib/parquet/AbstractParquetFileReader.java | 6 +-
.../contrib/parquet/ParquetFilePOJOReader.java | 6 +-
.../malhar/contrib/parser/AbstractCsvParser.java | 66 +++--
.../contrib/parser/CellProcessorBuilder.java | 5 +-
.../malhar/contrib/parser/CommonLogParser.java | 48 ++--
.../apex/malhar/contrib/parser/CsvParser.java | 10 +-
.../malhar/contrib/parser/DelimitedSchema.java | 2 +-
.../malhar/contrib/parser/FixedWidthParser.java | 20 +-
.../malhar/contrib/parser/FixedWidthSchema.java | 17 +-
.../apex/malhar/contrib/parser/JsonParser.java | 6 +-
.../apex/malhar/contrib/parser/LogParser.java | 11 +-
.../malhar/contrib/parser/LogSchemaDetails.java | 18 +-
.../apex/malhar/contrib/parser/RegexParser.java | 10 +-
.../apache/apex/malhar/contrib/parser/Schema.java | 2 +-
.../malhar/contrib/parser/StreamingJsonParser.java | 10 +-
.../apex/malhar/contrib/parser/package-info.java | 2 +-
.../apex/malhar/contrib/r/REngineConnectable.java | 12 +-
.../org/apache/apex/malhar/contrib/r/RMax.java | 14 +-
.../org/apache/apex/malhar/contrib/r/RMin.java | 14 +-
.../org/apache/apex/malhar/contrib/r/RScript.java | 87 +++----
.../apex/malhar/contrib/r/RStandardDeviation.java | 12 +-
.../apache/apex/malhar/contrib/r/package-info.java | 2 +-
.../rabbitmq/AbstractRabbitMQInputOperator.java | 62 ++---
.../rabbitmq/AbstractRabbitMQOutputOperator.java | 48 ++--
.../AbstractSinglePortRabbitMQInputOperator.java | 15 +-
.../AbstractSinglePortRabbitMQOutputOperator.java | 5 +-
.../contrib/rabbitmq/RabbitMQInputOperator.java | 2 +-
.../contrib/rabbitmq/RabbitMQOutputOperator.java | 2 +-
.../apex/malhar/contrib/rabbitmq/package-info.java | 2 +-
.../AbstractRedisAggregateOutputOperator.java | 6 +-
.../contrib/redis/AbstractRedisInputOperator.java | 21 +-
.../apex/malhar/contrib/redis/LettuceStore.java | 13 +-
.../apex/malhar/contrib/redis/NumberSummation.java | 44 ++--
.../redis/RedisKeyValPairOutputOperator.java | 5 +-
.../contrib/redis/RedisKeyValueInputOperator.java | 4 +-
.../redis/RedisMapAsValueInputOperator.java | 4 +-
.../contrib/redis/RedisMapOutputOperator.java | 2 +-
...disNumberSummationKeyValPairOutputOperator.java | 5 +-
.../RedisNumberSummationMapOutputOperator.java | 2 +-
.../contrib/redis/RedisPOJOInputOperator.java | 38 +--
.../contrib/redis/RedisPOJOOutputOperator.java | 14 +-
.../apex/malhar/contrib/redis/RedisStore.java | 34 ++-
.../apex/malhar/contrib/redis/package-info.java | 2 +-
.../contrib/romesyndication/RomeFeedEntry.java | 6 +-
.../romesyndication/RomeStreamProvider.java | 2 +-
.../romesyndication/RomeSyndicationOperator.java | 30 +--
.../contrib/romesyndication/package-info.java | 2 +-
.../apex/malhar/contrib/ruby/RubyOperator.java | 58 ++---
.../contrib/solr/AbstractSolrInputOperator.java | 8 +-
.../contrib/solr/AbstractSolrOutputOperator.java | 12 +-
.../contrib/solr/CloudSolrServerConnector.java | 2 +-
.../solr/ConcurrentUpdateSolrServerConnector.java | 2 +-
.../contrib/solr/HttpSolrServerConnector.java | 2 +-
.../contrib/solr/LBHttpSolrServerConnector.java | 2 +-
.../malhar/contrib/solr/SolrInputOperator.java | 6 +-
.../malhar/contrib/solr/SolrOutputOperator.java | 2 +-
.../malhar/contrib/solr/SolrServerConnector.java | 5 +-
.../apex/malhar/contrib/solr/package-info.java | 2 +-
.../splunk/AbstractSplunkInputOperator.java | 24 +-
.../contrib/splunk/SplunkInputFromForwarder.java | 56 +++--
.../malhar/contrib/splunk/SplunkInputOperator.java | 2 +-
.../apex/malhar/contrib/splunk/SplunkStore.java | 38 +--
.../contrib/splunk/SplunkTcpOutputOperator.java | 38 +--
.../apex/malhar/contrib/splunk/package-info.java | 2 +-
.../contrib/sqlite/SqliteStreamOperator.java | 27 +--
.../apex/malhar/contrib/sqlite/package-info.java | 2 +-
.../malhar/contrib/twitter/TwitterSampleInput.java | 40 +--
.../zmq/AbstractBaseZeroMQInputOperator.java | 26 +-
.../zmq/AbstractBaseZeroMQOutputOperator.java | 10 +-
.../zmq/AbstractSinglePortZeroMQInputOperator.java | 15 +-
.../AbstractSinglePortZeroMQOutputOperator.java | 2 +-
.../SimpleSinglePortZeroMQPullInputOperator.java | 5 +-
...pleSinglePortZeroMQPullStringInputOperator.java | 2 +-
.../malhar/contrib/zmq/ZeroMQInputOperator.java | 5 +-
.../malhar/contrib/zmq/ZeroMQOutputOperator.java | 15 +-
.../apex/malhar/contrib/zmq/package-info.java | 2 +-
.../accumulo/AccumuloInputOperatorTest.java | 29 ++-
.../accumulo/AccumuloOutputOperatorTest.java | 30 +--
.../accumulo/AccumuloRowTupleGenerator.java | 21 +-
.../contrib/accumulo/AccumuloTestHelper.java | 39 +--
.../malhar/contrib/accumulo/AccumuloTuple.java | 31 ++-
.../contrib/accumulo/AccumuloTupleCollector.java | 18 +-
.../contrib/aerospike/AerospikeOperatorTest.java | 93 +++----
...rospikePOJONonTransactionalPutOperatorTest.java | 23 +-
.../AerospikePOJOTransactionalPutOperatorTest.java | 23 +-
.../contrib/aerospike/AerospikeTestUtils.java | 123 +++++-----
.../apachelog/ApacheLogInputGeneratorTest.java | 6 +-
.../contrib/apachelog/ApplicationLocalLogTest.java | 18 +-
.../apachelog/ApplicationLogGeneratorTest.java | 18 +-
.../contrib/apachelog/zmq/ApplicationTest.java | 16 +-
.../contrib/avro/AvroFileInputOperatorTest.java | 27 ++-
.../contrib/avro/AvroFileToPojoModuleTest.java | 8 +-
.../apex/malhar/contrib/avro/AvroToPojoTest.java | 6 +-
.../apex/malhar/contrib/avro/PojoToAvroTest.java | 7 +-
.../AbstractUpsertOutputOperatorCodecsTest.java | 7 +-
...bstractUpsertOutputOperatorCompositePKTest.java | 7 +-
.../AbstractUpsertOutputOperatorCountersTest.java | 8 +-
.../apex/malhar/contrib/cassandra/Address.java | 2 +-
.../malhar/contrib/cassandra/AddressCodec.java | 2 +-
.../CassandraLookupCacheBackedOperatorTest.java | 16 +-
.../contrib/cassandra/CassandraOperatorTest.java | 33 ++-
.../contrib/cassandra/CompositePrimaryKeyRow.java | 2 +-
.../CompositePrimaryKeyUpdateOperator.java | 3 +-
.../contrib/cassandra/CounterColumnTableEntry.java | 2 +-
.../cassandra/CounterColumnUpdatesOperator.java | 3 +-
.../apex/malhar/contrib/cassandra/FullName.java | 2 +-
.../malhar/contrib/cassandra/FullNameCodec.java | 2 +-
.../malhar/contrib/cassandra/TestInputPojo.java | 3 +-
.../apache/apex/malhar/contrib/cassandra/User.java | 2 +-
.../contrib/cassandra/UserUpsertOperator.java | 3 +-
.../malhar/contrib/couchbase/CouchBaseGetTest.java | 20 +-
.../couchbase/CouchBaseInputOperatorTest.java | 30 ++-
.../couchbase/CouchBaseOutputOperatorTest.java | 28 +--
.../contrib/couchbase/CouchBasePOJOTest.java | 39 ++-
.../malhar/contrib/couchbase/CouchBaseSetTest.java | 17 +-
.../contrib/couchbase/TestComplexPojoInput.java | 2 +-
.../apex/malhar/contrib/couchbase/TestPojo.java | 3 +-
.../contrib/couchdb/CouchDBInputOperatorTest.java | 14 +-
.../contrib/couchdb/CouchDBOutputOperatorTest.java | 13 +-
.../couchdb/CouchDBPOJOInputOperatorTest.java | 8 +-
.../malhar/contrib/couchdb/CouchDBTestHelper.java | 8 +-
.../couchdb/PagedCouchDBInputOperatorTest.java | 11 +-
.../apex/malhar/contrib/couchdb/TestInputPOJO.java | 2 +-
.../elasticsearch/ElasticSearchOperatorTest.java | 16 +-
.../elasticsearch/ElasticSearchPercolateTest.java | 32 ++-
.../apex/malhar/contrib/enrich/EmployeeOrder.java | 2 +-
.../malhar/contrib/enrich/FileEnrichmentTest.java | 7 +-
.../apex/malhar/contrib/enrich/JDBCLoaderTest.java | 5 +-
.../malhar/contrib/enrich/MapEnricherTest.java | 10 +-
.../apache/apex/malhar/contrib/enrich/Order.java | 2 +-
.../malhar/contrib/enrich/POJOEnricherTest.java | 6 +-
.../malhar/contrib/formatter/CsvFormatterTest.java | 12 +-
.../contrib/geode/GeodeCheckpointStoreTest.java | 4 +-
.../geode/GeodeKeyValueStorageAgentTest.java | 4 +-
.../malhar/contrib/geode/GeodeOperatorTest.java | 4 +-
.../contrib/geode/GeodePOJOOperatorTest.java | 9 +-
.../apex/malhar/contrib/geode/GeodeStoreTest.java | 2 +-
.../contrib/hbase/HBaseColTupleGenerator.java | 6 +-
.../hbase/HBaseCsvMappingPutOperatorTest.java | 12 +-
.../malhar/contrib/hbase/HBaseGetOperatorTest.java | 9 +-
.../hbase/HBaseNameValueCsvPutOperatorTest.java | 9 +-
.../contrib/hbase/HBasePOJOInputOperatorTest.java | 45 ++--
.../contrib/hbase/HBasePOJOPutOperatorTest.java | 134 +++++-----
.../contrib/hbase/HBaseRowStringGenerator.java | 20 +-
.../contrib/hbase/HBaseRowTupleGenerator.java | 6 +-
.../contrib/hbase/HBaseScanOperatorTest.java | 10 +-
.../apex/malhar/contrib/hbase/HBaseTestHelper.java | 70 +++---
.../HBaseTransactionalAppendOperatorTest.java | 41 ++--
.../hbase/HBaseTransactionalPutOperatorTest.java | 79 +++---
.../apex/malhar/contrib/hbase/HBaseTuple.java | 2 +-
.../malhar/contrib/hbase/HBaseTupleCollector.java | 27 +--
.../apex/malhar/contrib/hbase/HBaseUtil.java | 30 +--
.../malhar/contrib/helper/CollectorModule.java | 8 +-
.../contrib/helper/MessageQueueTestHelper.java | 11 +-
.../apex/malhar/contrib/helper/SourceModule.java | 6 +-
.../malhar/contrib/jython/PythonOperatorTest.java | 6 +-
.../kafka/KafkaExactlyOnceOutputOperatorTest.java | 16 +-
.../contrib/kafka/KafkaInputOperatorTest.java | 30 +--
.../contrib/kafka/KafkaOperatorTestBase.java | 28 ++-
.../contrib/kafka/KafkaOutputOperatorTest.java | 22 +-
.../kafka/KafkaPartitionableInputOperatorTest.java | 24 +-
.../malhar/contrib/kafka/KafkaSimpleConsumer.java | 11 +-
.../malhar/contrib/kafka/KafkaTestConsumer.java | 12 +-
.../malhar/contrib/kafka/KafkaTestPartitioner.java | 9 +-
.../malhar/contrib/kafka/KafkaTestProducer.java | 25 +-
.../malhar/contrib/kafka/OffsetManagerTest.java | 49 ++--
.../contrib/kafka/SimpleKakfaConsumerTest.java | 5 +-
.../KinesisByteArrayOutputOperatorTest.java | 23 +-
.../contrib/kinesis/KinesisInputOperatorTest.java | 8 +-
.../contrib/kinesis/KinesisOperatorTestBase.java | 30 +--
.../contrib/kinesis/KinesisOutputOperatorTest.java | 48 ++--
.../KinesisPartitionableInputOperatorTest.java | 13 +-
.../kinesis/KinesisStringOutputOperatorTest.java | 7 +-
.../contrib/kinesis/KinesisTestConsumer.java | 44 ++--
.../contrib/kinesis/KinesisTestProducer.java | 21 +-
.../malhar/contrib/kinesis/ShardManagerTest.java | 57 ++---
.../contrib/memcache/MemcacheOperatorTest.java | 10 +-
.../contrib/memcache/MemcachePOJOOperatorTest.java | 24 +-
.../MemcacheWhalinOperatorTest.java | 4 +-
.../memsql/AbstractMemsqlInputOperatorTest.java | 57 +++--
.../memsql/AbstractMemsqlOutputOperatorTest.java | 55 ++---
.../malhar/contrib/memsql/MemsqlInputOperator.java | 12 +-
.../malhar/contrib/memsql/MemsqlStoreTest.java | 2 +-
.../apex/malhar/contrib/memsql/TestInputPojo.java | 2 +-
.../algo/AbstractStreamPatternMatcherTest.java | 3 +-
.../contrib/misc/algo/AllAfterMatchMapTest.java | 2 +-
.../malhar/contrib/misc/algo/DistinctMapTest.java | 2 +-
.../contrib/misc/algo/FilterKeyValsTest.java | 2 +-
.../contrib/misc/algo/FilterKeysHashMapTest.java | 2 +-
.../contrib/misc/algo/FilterKeysMapTest.java | 2 +-
.../contrib/misc/algo/FirstMatchMapTest.java | 2 +-
.../apex/malhar/contrib/misc/algo/FirstNTest.java | 2 +-
.../contrib/misc/algo/FirstTillMatchTest.java | 2 +-
.../contrib/misc/algo/InsertSortDescTest.java | 2 +-
.../contrib/misc/algo/InvertIndexArrayTest.java | 4 +-
.../malhar/contrib/misc/algo/InvertIndexTest.java | 4 +-
.../malhar/contrib/misc/algo/LastMatchMapTest.java | 2 +-
.../contrib/misc/algo/LeastFrequentKeyMapTest.java | 2 +-
.../misc/algo/LeastFrequentKeyValueMapTest.java | 2 +-
.../malhar/contrib/misc/algo/MatchMapTest.java | 6 +-
.../contrib/misc/algo/MostFrequentKeyMapTest.java | 2 +-
.../misc/algo/MostFrequentKeyValueMapTest.java | 2 +-
.../apex/malhar/contrib/misc/algo/SamplerTest.java | 2 +-
.../contrib/misc/math/ChangeAlertKeyValTest.java | 4 +-
.../contrib/misc/math/ChangeAlertMapTest.java | 2 +-
.../malhar/contrib/misc/math/ChangeAlertTest.java | 4 +-
.../malhar/contrib/misc/math/ChangeKeyValTest.java | 4 +-
.../apex/malhar/contrib/misc/math/ChangeTest.java | 2 +-
.../contrib/misc/math/CompareExceptMapTest.java | 2 +-
.../malhar/contrib/misc/math/CompareMapTest.java | 2 +-
.../malhar/contrib/misc/math/CountKeyValTest.java | 4 +-
.../malhar/contrib/misc/math/ExceptMapTest.java | 2 +-
.../malhar/contrib/misc/math/QuotientMapTest.java | 2 +-
.../malhar/contrib/misc/math/SumCountMapTest.java | 2 +-
.../misc/streamquery/DeleteOperatorTest.java | 3 +-
.../streamquery/FullOuterJoinOperatorTest.java | 9 +-
.../misc/streamquery/GroupByOperatorTest.java | 5 +-
.../misc/streamquery/HavingOperatorTest.java | 5 +-
.../misc/streamquery/InnerJoinOperatorTest.java | 10 +-
.../streamquery/LeftOuterJoinOperatorTest.java | 8 +-
.../misc/streamquery/OrderByOperatorTest.java | 4 +-
.../streamquery/RightOuterJoinOperatorTest.java | 8 +-
.../misc/streamquery/SelectOperatorTest.java | 5 +-
.../misc/streamquery/SelectTopOperatorTest.java | 2 +-
.../misc/streamquery/UpdateOperatorTest.java | 3 +-
.../streamquery/advanced/BetweenConditionTest.java | 5 +-
.../advanced/CompoundConditionTest.java | 5 +-
.../misc/streamquery/advanced/InConditionTest.java | 5 +-
.../streamquery/advanced/LikeConditionTest.java | 5 +-
.../misc/streamquery/advanced/NegateIndexTest.java | 3 +-
.../streamquery/advanced/SelectAverageTest.java | 3 +-
.../misc/streamquery/advanced/SelectCountTest.java | 3 +-
.../streamquery/advanced/SelectFirstLastTest.java | 3 +-
.../streamquery/advanced/SelectMaxMinTest.java | 3 +-
.../misc/streamquery/advanced/SumIndexTest.java | 3 +-
.../contrib/mongodb/MongoDBInputOperatorTest.java | 20 +-
.../contrib/mongodb/MongoDBOutputOperatorTest.java | 32 ++-
.../apex/malhar/contrib/mongodb/TestPOJO.java | 2 +-
.../malhar/contrib/mqtt/MqttInputOperatorTest.java | 31 +--
.../contrib/mqtt/MqttOutputOperatorTest.java | 30 +--
.../nifi/NiFiSinglePortInputOperatorTest.java | 15 +-
.../nifi/NiFiSinglePortOutputOperatorTest.java | 29 +--
.../nifi/demo/TestNiFiInputApplication.java | 6 +-
.../nifi/demo/TestNiFiOutputApplication.java | 17 +-
.../malhar/contrib/nifi/mock/MockDataPacket.java | 2 +-
.../contrib/nifi/mock/MockSiteToSiteClient.java | 2 +-
.../malhar/contrib/nifi/mock/MockTransaction.java | 2 +-
.../contrib/parquet/ParquetFilePOJOReaderTest.java | 37 ++-
.../apex/malhar/contrib/parser/CSVParserTest.java | 34 +--
.../malhar/contrib/parser/CommonLogParserTest.java | 3 +-
.../malhar/contrib/parser/CsvPOJOParserTest.java | 8 +-
.../apex/malhar/contrib/parser/FixedWidthTest.java | 10 +-
.../contrib/parser/JsonParserApplicationTest.java | 8 +-
.../apex/malhar/contrib/parser/JsonParserTest.java | 10 +-
.../apex/malhar/contrib/parser/LogParserTest.java | 36 +--
.../malhar/contrib/parser/RegexParserTest.java | 38 ++-
.../contrib/parser/StreamingJsonParserTest.java | 8 +-
.../apex/malhar/contrib/r/RMaxOperatorTest.java | 6 +-
.../apex/malhar/contrib/r/RMinOperatorTest.java | 6 +-
.../r/RScriptLinearRegressionPredictorTest.java | 14 +-
.../contrib/r/RScriptOperatorBooleanTest.java | 16 +-
.../contrib/r/RScriptOperatorDoubleArrayTest.java | 10 +-
.../malhar/contrib/r/RScriptOperatorIntTest.java | 6 +-
.../malhar/contrib/r/RScriptOperatorRealTest.java | 8 +-
.../malhar/contrib/r/RScriptOperatorStrTest.java | 4 +-
.../contrib/r/RStandardDeviationOperatorTest.java | 8 +-
.../rabbitmq/RabbitMQInputOperatorBenchmark.java | 2 +-
.../rabbitmq/RabbitMQInputOperatorTest.java | 22 +-
.../rabbitmq/RabbitMQOutputOperatorBenchmark.java | 2 +-
.../rabbitmq/RabbitMQOutputOperatorTest.java | 14 +-
.../contrib/redis/RedisInputOperatorTest.java | 15 +-
.../malhar/contrib/redis/RedisOperatorTest.java | 4 +-
.../contrib/redis/RedisPOJOOperatorTest.java | 20 +-
.../RomeSyndicationOperatorTest.java | 25 +-
.../apex/malhar/contrib/ruby/RubyOperatorTest.java | 18 +-
.../splunk/SplunkInputFromForwarderTest.java | 17 +-
.../contrib/splunk/SplunkInputOperatorTest.java | 11 +-
.../splunk/SplunkTcpOutputOperatorTest.java | 32 +--
.../util/FieldValueSerializableGenerator.java | 50 ++--
.../contrib/util/POJOTupleGenerateOperator.java | 36 ++-
.../apache/apex/malhar/contrib/util/TestPOJO.java | 64 ++---
.../contrib/util/TupleCacheOutputOperator.java | 13 +-
.../contrib/util/TupleGenerateCacheOperator.java | 9 +-
.../apex/malhar/contrib/util/TupleGenerator.java | 33 +--
.../contrib/zmq/ZeroMQInputOperatorBenchmark.java | 5 +-
.../contrib/zmq/ZeroMQInputOperatorTest.java | 21 +-
.../malhar/contrib/zmq/ZeroMQMessageGenerator.java | 12 +-
.../malhar/contrib/zmq/ZeroMQMessageReceiver.java | 13 +-
.../contrib/zmq/ZeroMQOutputOperatorBenchmark.java | 2 +-
.../contrib/zmq/ZeroMQOutputOperatorTest.java | 21 +-
examples/csvformatter/pom.xml | 2 +-
.../apex/examples/csvformatter/Application.java | 6 +-
.../examples/csvformatter/HDFSOutputOperator.java | 3 +-
examples/dedup/pom.xml | 2 +-
.../apache/apex/examples/dedup/Application.java | 2 +-
examples/distributedistinct/pom.xml | 2 +-
.../examples/distributeddistinct/Application.java | 10 +-
.../distributeddistinct/CountVerifier.java | 3 +-
.../IntegerUniqueValueCountAppender.java | 4 +-
.../distributeddistinct/RandomKeyValGenerator.java | 5 +-
.../distributeddistinct/StatefulApplication.java | 8 +-
.../UniqueValueCountAppender.java | 7 +-
.../DistributedDistinctTest.java | 8 +-
.../StatefulUniqueCountTest.java | 5 +-
examples/dynamic-partition/pom.xml | 2 +-
.../java/org/apache/apex/examples/dynamic/App.java | 3 +-
examples/echoserver/pom.xml | 2 +-
examples/enricher/pom.xml | 2 +-
.../examples/enricher/EnricherAppWithJSONFile.java | 6 +-
.../apex/examples/enricher/LineOutputOperator.java | 2 +-
examples/exactly-once/pom.xml | 2 +-
.../exactlyonce/ExactlyOnceFileOutputApp.java | 6 +-
.../exactlyonce/ExactlyOnceJdbcOutputApp.java | 10 +-
.../exactlyonce/ExactlyOnceJdbcOutputTest.java | 2 +-
examples/filter/pom.xml | 2 +-
.../apache/apex/examples/filter/Application.java | 6 +-
examples/frauddetect/pom.xml | 2 +-
.../apex/examples/frauddetect/Application.java | 14 +-
.../frauddetect/AverageAlertingOperator.java | 3 +-
.../apex/examples/frauddetect/BankIdNumberKey.java | 2 +-
.../frauddetect/BankIdNumberSamplerOperator.java | 4 +-
.../CreditCardAmountSamplerOperator.java | 2 +-
.../MerchantTransactionBucketOperator.java | 2 +-
.../frauddetect/SlidingWindowSumKeyVal.java | 6 +-
.../frauddetect/SlidingWindowSumObject.java | 3 +-
.../frauddetect/TransactionStatsAggregator.java | 4 +-
.../operator/HdfsStringOutputOperator.java | 3 +-
examples/ftp/pom.xml | 2 +-
.../org/apache/apex/examples/ftp/Application.java | 2 +-
examples/highlevelapi/pom.xml | 2 +-
.../malhar/stream/sample/MinimalWordCount.java | 2 +-
.../malhar/stream/sample/WindowedWordCount.java | 2 +-
.../stream/sample/complete/AutoComplete.java | 2 +-
.../sample/complete/StreamingWordExtract.java | 6 +-
.../sample/complete/TopWikipediaSessions.java | 2 +-
.../stream/sample/complete/TrafficRoutes.java | 2 +-
.../sample/complete/TwitterAutoComplete.java | 4 +-
.../sample/cookbook/CombinePerKeyExamples.java | 2 +-
.../stream/sample/cookbook/MaxPerKeyExamples.java | 14 +-
.../stream/sample/cookbook/TriggerExample.java | 3 +-
.../stream/sample/WindowedWordCountTest.java | 2 +-
.../sample/complete/StreamingWordExtractTest.java | 2 +-
.../stream/sample/complete/TrafficRoutesTest.java | 2 +-
.../sample/cookbook/MaxPerKeyExamplesTest.java | 2 +-
examples/iteration/pom.xml | 2 +-
.../apex/examples/iteration/Application.java | 2 +-
examples/jdbc/pom.xml | 2 +-
.../FileToJdbcApp/FileToJdbcCsvParser.java | 10 +-
.../FileToJdbcApp/FileToJdbcCustomParser.java | 6 +-
.../JdbcIngest/FileLineOutputOperator.java | 2 +-
.../apex/examples/JdbcIngest/JdbcHDFSApp.java | 8 +-
.../examples/JdbcIngest/JdbcPollerApplication.java | 8 +-
.../apex/examples/JdbcToJdbc/JdbcToJdbcApp.java | 22 +-
.../examples/FileToJdbcApp/ApplicationTest.java | 2 +-
.../apex/examples/JdbcToJdbc/JdbcOperatorTest.java | 6 +-
examples/kafka/pom.xml | 2 +-
.../kafka/exactlyonceoutput/ValidationToFile.java | 3 +-
.../examples/kafka/hdfs2kafka/Application.java | 2 +-
.../kafka/kafka2hdfs/LineOutputOperator.java | 2 +-
examples/machinedata/pom.xml | 2 +-
.../apex/examples/machinedata/Application.java | 4 +-
.../machinedata/operator/CalculatorOperator.java | 5 +-
.../operator/MachineInfoAveragingOperator.java | 5 +-
.../MachineInfoAveragingPrerequisitesOperator.java | 2 +-
.../operator/MachineInfoAveragingUnifier.java | 3 +-
.../machinedata/CalculatorOperatorTest.java | 8 +-
examples/mobile/pom.xml | 2 +-
.../apache/apex/examples/mobile/Application.java | 10 +-
.../examples/mobile/PhoneMovementGenerator.java | 4 +-
.../apex/examples/mobile/ApplicationTest.java | 10 +-
examples/mrmonitor/pom.xml | 2 +-
.../apex/examples/mrmonitor/Application.java | 4 +-
.../mrmonitor/MRMonitoringApplication.java | 5 +-
.../mrmonitor/MrMonitoringApplicationTest.java | 2 +-
examples/mroperator/pom.xml | 2 +-
.../mroperator/HdfsKeyValOutputOperator.java | 4 +-
.../apex/examples/mroperator/MapOperator.java | 3 +-
.../examples/mroperator/OutputCollectorImpl.java | 2 +-
.../apex/examples/mroperator/ReduceOperator.java | 2 +-
.../apex/examples/mroperator/MapOperatorTest.java | 3 +-
.../examples/mroperator/ReduceOperatorTest.java | 5 +-
examples/nyctaxi/pom.xml | 2 +-
.../apache/apex/examples/nyctaxi/Application.java | 6 +-
.../apex/examples/nyctaxi/NycTaxiDataServer.java | 2 +-
.../examples/nyctaxi/NycTaxiZipFareExtractor.java | 2 +-
examples/partition/pom.xml | 2 +-
.../org/apache/apex/examples/partition/Codec3.java | 2 +-
examples/pi/pom.xml | 2 +-
.../org/apache/apex/examples/pi/Application.java | 6 +-
.../apex/examples/pi/ApplicationAppData.java | 14 +-
.../apex/examples/pi/ApplicationWithScript.java | 14 +-
.../org/apache/apex/examples/pi/Calculator.java | 22 +-
.../pi/src/main/resources/app/PiJsonExample.json | 4 +-
examples/pom.xml | 2 +-
examples/r/pom.xml | 2 +-
.../examples/r/oldfaithful/FaithfulRScript.java | 3 +-
.../r/oldfaithful/OldFaithfulApplication.java | 2 +-
examples/recordReader/pom.xml | 2 +-
.../apex/examples/recordReader/Application.java | 4 +-
examples/s3/pom.xml | 2 +-
.../apache/apex/examples/s3Output/Application.java | 2 +-
.../s3ToHdfsSync/S3ToHDFSSyncApplication.java | 4 +-
.../META-INF/properties-s3TupleOutput.xml | 2 +-
examples/sql/pom.xml | 2 +-
.../sql/sample/FusionStyleSQLApplication.java | 2 +-
examples/throttle/pom.xml | 2 +-
examples/transform/pom.xml | 2 +-
.../transform/DynamicTransformApplication.java | 4 +-
.../transform/SimpleTransformApplication.java | 2 +-
examples/twitter/pom.xml | 2 +-
.../twitter/KinesisHashtagsApplication.java | 24 +-
.../examples/twitter/TwitterDumpApplication.java | 5 +-
.../twitter/TwitterDumpHBaseApplication.java | 5 +-
.../twitter/TwitterTopCounterApplication.java | 19 +-
.../twitter/TwitterTopWordsApplication.java | 4 +-
.../TwitterTrendingHashtagsApplication.java | 8 +-
.../examples/twitter/TwitterTopCounterTest.java | 3 +-
.../apex/examples/twitter/TwitterTopWordsTest.java | 3 +-
examples/uniquecount/pom.xml | 2 +-
.../apex/examples/uniquecount/Application.java | 12 +-
.../apex/examples/uniquecount/CountVerifier.java | 3 +-
.../examples/uniquecount/RandomDataGenerator.java | 3 +-
.../apex/examples/uniquecount/RandomKeyValues.java | 3 +-
.../examples/uniquecount/RandomKeysGenerator.java | 2 +-
.../uniquecount/UniqueKeyValCountExample.java | 9 +-
examples/wordcount/pom.xml | 2 +-
.../apex/examples/wordcount/Application.java | 8 +-
.../wordcount/ApplicationWithQuerySupport.java | 11 +-
.../apache/apex/examples/wordcount/LineReader.java | 2 +-
.../examples/wordcount/WordCountInputOperator.java | 2 +-
.../apex/examples/wordcount/WordCountWriter.java | 2 +-
examples/yahoofinance/pom.xml | 2 +-
.../yahoofinance/ApplicationWithDerbySQL.java | 2 +-
.../apex/examples/yahoofinance/StockTickInput.java | 2 +-
.../yahoofinance/YahooFinanceApplication.java | 40 +--
.../yahoofinance/YahooFinanceCSVInputOperator.java | 2 +-
.../src/test/resources/alert_create.json | 6 +-
flume/pom.xml | 2 +-
.../apex/malhar/flume/storage/HDFSStorage.java | 3 +-
hive/pom.xml | 2 +-
.../hive/AbstractFSRollingOutputOperator.java | 4 +-
.../apex/malhar/hive/FSPojoToHiveOperator.java | 22 +-
.../org/apache/apex/malhar/hive/HiveOperator.java | 8 +-
.../apache/apex/malhar/hive/HiveOutputModule.java | 5 +-
.../org/apache/apex/malhar/hive/HiveStore.java | 4 +-
.../apache/apex/malhar/hive/HiveStreamCodec.java | 6 +-
.../org/apache/apex/malhar/hive/package-info.java | 2 +-
.../apex/malhar/hive/FSRollingMapTestImpl.java | 2 +-
.../apache/apex/malhar/hive/FSRollingTestImpl.java | 2 +-
.../org/apache/apex/malhar/hive/HiveMockTest.java | 10 +-
.../org/apache/apex/malhar/hive/HiveStoreTest.java | 2 +-
kafka/kafka-common/pom.xml | 2 +-
.../malhar/kafka/AbstractKafkaPartitioner.java | 2 +-
kafka/kafka010/pom.xml | 2 +-
.../apex/malhar/kafka/KafkaOperatorTestBase.java | 2 +-
kafka/kafka09/pom.xml | 2 +-
.../apex/malhar/kafka/KafkaOperatorTestBase.java | 2 +-
.../apex/malhar/kafka/KafkaOutputOperatorTest.java | 2 +-
kafka/pom.xml | 2 +-
kudu/pom.xml | 2 +-
.../malhar/kudu/AbstractKuduInputOperator.java | 2 +-
.../malhar/kudu/AbstractKuduOutputOperator.java | 2 +-
.../apex/malhar/kudu/ApexKuduConnection.java | 7 +-
.../partitioner/AbstractKuduInputPartitioner.java | 2 +-
.../kudu/scanner/KuduPartitionScannerCallable.java | 3 +-
.../kudu/IncrementalStepScanInputOperatorTest.java | 4 +-
.../apex/malhar/kudu/KuduClientTestCommons.java | 3 +-
.../KuduCreateUpdateDeleteOutputOperatorTest.java | 4 +-
.../apex/malhar/kudu/KuduInputOperatorCommons.java | 4 +-
.../partitioner/KuduOneToManyPartitionerTest.java | 2 +-
.../partitioner/KuduOneToOnePartitionerTest.java | 2 +-
library/pom.xml | 2 +-
.../apache/apex/malhar/lib/algo/BottomNMap.java | 6 +-
.../apex/malhar/lib/algo/BottomNUnifier.java | 5 +-
.../apex/malhar/lib/algo/BottomNUniqueMap.java | 6 +-
.../malhar/lib/algo/CompareExceptCountMap.java | 6 +-
.../org/apache/apex/malhar/lib/algo/Distinct.java | 6 +-
.../apache/apex/malhar/lib/algo/FilterValues.java | 2 +-
.../apache/apex/malhar/lib/algo/InsertSort.java | 6 +-
.../lib/algo/LeastFrequentKeyArrayUnifier.java | 2 +-
.../malhar/lib/algo/LeastFrequentKeyUnifier.java | 2 +-
.../apex/malhar/lib/algo/LeastFrequentValue.java | 6 +-
.../apache/apex/malhar/lib/algo/MatchAllMap.java | 8 +-
.../apache/apex/malhar/lib/algo/MatchAnyMap.java | 8 +-
.../org/apache/apex/malhar/lib/algo/MatchMap.java | 8 +-
.../org/apache/apex/malhar/lib/algo/MergeSort.java | 2 +-
.../apex/malhar/lib/algo/MergeSortNumber.java | 2 +-
.../apex/malhar/lib/algo/MostFrequentValue.java | 10 +-
.../java/org/apache/apex/malhar/lib/algo/TopN.java | 6 +-
.../apache/apex/malhar/lib/algo/TopNUnique.java | 6 +-
.../apache/apex/malhar/lib/algo/UniqueCounter.java | 7 +-
.../apex/malhar/lib/algo/UniqueValueCount.java | 5 +-
.../apache/apex/malhar/lib/algo/package-info.java | 2 +-
.../malhar/lib/appdata/AbstractAppDataServer.java | 1 -
.../apache/apex/malhar/lib/appdata/QueueUtils.java | 2 +-
.../apache/apex/malhar/lib/appdata/StoreUtils.java | 5 +-
.../datastructs/CacheLRUSynchronousFlush.java | 2 +-
.../lib/appdata/datastructs/DimensionalTable.java | 2 +-
.../appdata/datastructs/NOPCacheFlushListener.java | 4 +-
.../malhar/lib/appdata/gpo/GPOByteArrayList.java | 2 +-
.../apex/malhar/lib/appdata/gpo/GPOGetters.java | 20 +-
.../apex/malhar/lib/appdata/gpo/GPOMutable.java | 10 +-
.../apex/malhar/lib/appdata/gpo/GPOType.java | 61 +++--
.../apex/malhar/lib/appdata/gpo/GPOUtils.java | 31 ++-
.../apache/apex/malhar/lib/appdata/gpo/Serde.java | 2 +-
.../lib/appdata/gpo/SerdeFieldsDescriptor.java | 7 +-
.../lib/appdata/gpo/SerdeListGPOMutable.java | 5 +-
.../malhar/lib/appdata/gpo/SerdeListPrimitive.java | 5 +-
.../malhar/lib/appdata/gpo/SerdeListString.java | 2 +-
.../malhar/lib/appdata/gpo/SerdeMapPrimitive.java | 5 +-
.../lib/appdata/gpo/SerdeObjectPayloadFix.java | 2 +-
.../apex/malhar/lib/appdata/package-info.java | 2 +-
.../query/AbstractWindowEndQueueManager.java | 7 +-
.../query/AppDataWindowEndQueueManager.java | 7 +-
.../apex/malhar/lib/appdata/query/QueryBundle.java | 2 +-
.../malhar/lib/appdata/query/QueryExecutor.java | 2 +-
.../appdata/query/QueryManagerAsynchronous.java | 7 +-
.../lib/appdata/query/QueryManagerSynchronous.java | 2 +-
.../apex/malhar/lib/appdata/query/QueueList.java | 2 +-
.../malhar/lib/appdata/query/QueueManager.java | 2 +-
.../lib/appdata/query/SimpleDoneQueueManager.java | 5 +-
.../lib/appdata/query/SimpleQueueManager.java | 5 +-
.../lib/appdata/query/WindowBoundedService.java | 2 +-
.../lib/appdata/query/WindowEndQueueManager.java | 5 +-
.../query/serde/CustomMessageDeserializer.java | 4 +-
.../query/serde/CustomMessageSerializer.java | 6 +-
.../query/serde/CustomMessageValidator.java | 4 +-
.../query/serde/DataQuerySnapshotDeserializer.java | 12 +-
.../query/serde/DataQuerySnapshotValidator.java | 12 +-
.../query/serde/DataResultSnapshotSerializer.java | 16 +-
.../query/serde/MessageDeserializerFactory.java | 6 +-
.../query/serde/MessageDeserializerInfo.java | 2 +-
.../query/serde/MessageSerializerFactory.java | 8 +-
.../appdata/query/serde/MessageSerializerInfo.java | 2 +-
.../lib/appdata/query/serde/MessageType.java | 2 +-
.../appdata/query/serde/MessageValidatorInfo.java | 2 +-
.../query/serde/SchemaQueryDeserializer.java | 10 +-
.../query/serde/SimpleDataDeserializer.java | 4 +-
.../appdata/query/serde/SimpleDataSerializer.java | 6 +-
.../appdata/query/serde/SimpleDataValidator.java | 4 +-
.../lib/appdata/schemas/CustomTimeBucket.java | 2 +-
.../lib/appdata/schemas/DataQuerySnapshot.java | 14 +-
.../lib/appdata/schemas/DataResultSnapshot.java | 12 +-
.../schemas/DimensionalConfigurationSchema.java | 4 +-
.../lib/appdata/schemas/DimensionalSchema.java | 2 +-
.../apex/malhar/lib/appdata/schemas/Fields.java | 2 +-
.../lib/appdata/schemas/FieldsDescriptor.java | 8 +-
.../apex/malhar/lib/appdata/schemas/JSONType.java | 2 +-
.../apex/malhar/lib/appdata/schemas/Message.java | 2 +-
.../apex/malhar/lib/appdata/schemas/QRBase.java | 2 +-
.../apex/malhar/lib/appdata/schemas/Query.java | 2 +-
.../apex/malhar/lib/appdata/schemas/Result.java | 2 +-
.../lib/appdata/schemas/ResultFormatter.java | 2 +-
.../apex/malhar/lib/appdata/schemas/Schema.java | 2 +-
.../malhar/lib/appdata/schemas/SchemaQuery.java | 12 +-
.../malhar/lib/appdata/schemas/SchemaRegistry.java | 2 +-
.../appdata/schemas/SchemaRegistryMultiple.java | 8 +-
.../lib/appdata/schemas/SchemaRegistrySingle.java | 2 +-
.../malhar/lib/appdata/schemas/SchemaResult.java | 8 +-
.../appdata/schemas/SchemaResultSerializer.java | 4 +-
.../malhar/lib/appdata/schemas/SchemaUtils.java | 2 +-
.../malhar/lib/appdata/schemas/SnapshotSchema.java | 2 +-
.../malhar/lib/appdata/schemas/TimeBucket.java | 2 +-
.../apex/malhar/lib/appdata/schemas/Type.java | 2 +-
.../snapshot/AbstractAppDataSnapshotServer.java | 36 +--
.../appdata/snapshot/AppDataSnapshotServerMap.java | 8 +-
.../snapshot/AppDataSnapshotServerPOJO.java | 8 +-
.../lib/bandwidth/BandwidthLimitingOperator.java | 2 +-
.../malhar/lib/bandwidth/BandwidthManager.java | 2 +-
.../malhar/lib/bandwidth/BandwidthPartitioner.java | 2 +-
.../lib/codec/JavaSerializationStreamCodec.java | 2 +-
.../apex/malhar/lib/codec/KryoJdkContainer.java | 2 +-
.../lib/codec/KryoSerializableStreamCodec.java | 2 +-
.../apache/apex/malhar/lib/codec/package-info.java | 2 +-
.../ByteArrayToStringConverterOperator.java | 2 +-
.../apex/malhar/lib/converter/Converter.java | 2 +-
.../converter/MapToKeyHashValuePairConverter.java | 5 +-
.../lib/converter/MapToKeyValuePairConverter.java | 5 +-
.../StringValueToNumberConverterForMap.java | 2 +-
.../apex/malhar/lib/converter/package-info.java | 2 +-
.../apex/malhar/lib/counters/BasicCounters.java | 2 +-
.../apex/malhar/lib/counters/package-info.java | 2 +-
...TransactionableKeyValueStoreOutputOperator.java | 2 +-
...ggregateTransactionableStoreOutputOperator.java | 2 +-
...actBatchTransactionableStoreOutputOperator.java | 2 +-
.../lib/db/AbstractKeyValueStoreInputOperator.java | 2 +-
...TransactionableKeyValueStoreOutputOperator.java | 2 +-
...PassThruTransactionableStoreOutputOperator.java | 2 +-
.../malhar/lib/db/AbstractStoreInputOperator.java | 2 +-
.../malhar/lib/db/AbstractStoreOutputOperator.java | 2 +-
...AbstractTransactionableStoreOutputOperator.java | 2 +-
.../org/apache/apex/malhar/lib/db/Connectable.java | 2 +-
.../apache/apex/malhar/lib/db/KeyValueStore.java | 2 +-
.../apache/apex/malhar/lib/db/Transactionable.java | 2 +-
.../lib/db/TransactionableKeyValueStore.java | 2 +-
.../apex/malhar/lib/db/TransactionableStore.java | 2 +-
.../cache/AbstractDBLookupCacheBackedOperator.java | 7 +-
.../apex/malhar/lib/db/cache/CacheManager.java | 5 +-
.../apex/malhar/lib/db/cache/CacheStore.java | 2 +-
.../apex/malhar/lib/db/cache/package-info.java | 2 +-
.../lib/db/jdbc/AbstractJdbcInputOperator.java | 6 +-
...tJdbcNonTransactionableBatchOutputOperator.java | 2 +-
...stractJdbcNonTransactionableOutputOperator.java | 5 +-
.../db/jdbc/AbstractJdbcPOJOOutputOperator.java | 21 +-
.../lib/db/jdbc/AbstractJdbcPollInputOperator.java | 8 +-
.../AbstractJdbcTransactionableOutputOperator.java | 5 +-
.../lib/db/jdbc/JDBCDimensionalOutputOperator.java | 12 +-
.../lib/db/jdbc/JDBCLookupCacheBackedOperator.java | 6 +-
.../apex/malhar/lib/db/jdbc/JdbcFieldInfo.java | 4 +-
.../lib/db/jdbc/JdbcNonTransactionalStore.java | 2 +-
.../malhar/lib/db/jdbc/JdbcPOJOInputOperator.java | 7 +-
.../lib/db/jdbc/JdbcPOJOInsertOutputOperator.java | 5 +-
.../db/jdbc/JdbcPOJONonInsertOutputOperator.java | 2 +-
.../lib/db/jdbc/JdbcPOJOPollInputOperator.java | 8 +-
.../malhar/lib/db/jdbc/JdbcPollInputOperator.java | 2 +-
.../apache/apex/malhar/lib/db/jdbc/JdbcStore.java | 5 +-
.../malhar/lib/db/jdbc/JdbcTransactionalStore.java | 4 +-
.../apex/malhar/lib/db/jdbc/package-info.java | 2 +-
.../apache/apex/malhar/lib/db/package-info.java | 2 +-
.../RedshiftJdbcTransactionableOutputOperator.java | 2 +-
.../lib/db/redshift/RedshiftOutputModule.java | 2 +-
.../apex/malhar/lib/dedup/AbstractDeduper.java | 2 +-
.../malhar/lib/dedup/BoundedDedupOperator.java | 5 +-
.../apex/malhar/lib/dedup/DeduperStreamCodec.java | 6 +-
.../malhar/lib/dedup/TimeBasedDedupOperator.java | 4 +-
.../lib/dimensions/CustomTimeBucketRegistry.java | 4 +-
.../dimensions/DimensionsConversionContext.java | 4 +-
.../lib/dimensions/DimensionsDescriptor.java | 14 +-
.../malhar/lib/dimensions/DimensionsEvent.java | 5 +-
.../aggregator/AbstractCompositeAggregator.java | 3 +-
.../aggregator/AbstractIncrementalAggregator.java | 9 +-
.../aggregator/AbstractTopBottomAggregator.java | 5 +-
.../dimensions/aggregator/AggregatorAverage.java | 9 +-
.../lib/dimensions/aggregator/AggregatorCount.java | 8 +-
.../dimensions/aggregator/AggregatorCumSum.java | 16 +-
.../lib/dimensions/aggregator/AggregatorFirst.java | 6 +-
.../lib/dimensions/aggregator/AggregatorLast.java | 6 +-
.../lib/dimensions/aggregator/AggregatorMax.java | 8 +-
.../lib/dimensions/aggregator/AggregatorMin.java | 8 +-
.../dimensions/aggregator/AggregatorRegistry.java | 2 +-
.../lib/dimensions/aggregator/AggregatorSum.java | 8 +-
.../lib/dimensions/aggregator/AggregatorUtils.java | 12 +-
.../dimensions/aggregator/CompositeAggregator.java | 5 +-
.../aggregator/IncrementalAggregator.java | 7 +-
.../lib/dimensions/aggregator/OTFAggregator.java | 4 +-
.../apex/malhar/lib/expression/Expression.java | 2 +-
.../lib/expression/JavaExpressionParser.java | 4 +-
.../apex/malhar/lib/fileaccess/DTFileReader.java | 2 +-
.../apex/malhar/lib/fileaccess/FileAccess.java | 2 +-
.../malhar/lib/fileaccess/FileAccessFSImpl.java | 2 +-
.../apex/malhar/lib/fileaccess/TFileImpl.java | 2 +-
.../apex/malhar/lib/fileaccess/TFileReader.java | 2 +-
.../apex/malhar/lib/fileaccess/TFileWriter.java | 2 +-
.../apex/malhar/lib/fileaccess/package-info.java | 2 +-
.../apex/malhar/lib/filter/FilterOperator.java | 6 +-
.../apex/malhar/lib/formatter/Formatter.java | 5 +-
.../apex/malhar/lib/formatter/JsonFormatter.java | 2 +-
.../apex/malhar/lib/formatter/XmlFormatter.java | 2 +-
.../apache/apex/malhar/lib/fs/FSRecordReader.java | 6 +-
.../apex/malhar/lib/fs/FSRecordReaderModule.java | 4 +-
.../malhar/lib/fs/GenericFileOutputOperator.java | 5 +-
.../malhar/lib/fs/LineByLineFileInputOperator.java | 3 +-
.../malhar/lib/fs/s3/S3BlockUploadOperator.java | 4 +-
.../lib/fs/s3/S3InitiateFileUploadOperator.java | 2 +-
.../apex/malhar/lib/fs/s3/S3OutputModule.java | 6 +-
.../apache/apex/malhar/lib/fs/s3/S3Reconciler.java | 2 +-
.../lib/fs/s3/S3ReconcilerQueuePartitioner.java | 3 +-
.../apex/malhar/lib/fs/s3/S3RecordReader.java | 5 +-
.../malhar/lib/fs/s3/S3RecordReaderModule.java | 3 +-
.../apex/malhar/lib/fs/s3/S3TupleOutputModule.java | 4 +-
.../apache/apex/malhar/lib/function/Function.java | 3 +-
.../malhar/lib/io/AbstractFTPInputOperator.java | 4 +-
.../io/AbstractHttpGetMultiValuedMapOperator.java | 2 +-
.../malhar/lib/io/AbstractHttpGetOperator.java | 2 +-
.../malhar/lib/io/AbstractHttpInputOperator.java | 2 +-
.../apex/malhar/lib/io/AbstractHttpOperator.java | 2 +-
.../io/AbstractKeyValueStoreOutputOperator.java | 5 +-
.../malhar/lib/io/AbstractSocketInputOperator.java | 2 +-
.../apex/malhar/lib/io/ApacheGenRandomLogs.java | 2 +-
.../io/CollectionMultiConsoleOutputOperator.java | 2 +-
.../apex/malhar/lib/io/ConsoleOutputOperator.java | 2 +-
.../apex/malhar/lib/io/HttpGetMapOperator.java | 2 +-
.../malhar/lib/io/HttpJsonChunksInputOperator.java | 2 +-
.../apex/malhar/lib/io/HttpLinesInputOperator.java | 2 +-
.../apex/malhar/lib/io/HttpOutputOperator.java | 2 +-
.../apex/malhar/lib/io/HttpPostOutputOperator.java | 2 +-
.../malhar/lib/io/IdempotentStorageManager.java | 4 +-
.../lib/io/MapMultiConsoleOutputOperator.java | 2 +-
.../malhar/lib/io/PubSubWebSocketAppDataQuery.java | 7 +-
.../lib/io/PubSubWebSocketAppDataResult.java | 2 +-
.../lib/io/PubSubWebSocketInputOperator.java | 2 +-
.../lib/io/PubSubWebSocketOutputOperator.java | 2 +-
.../lib/io/SimpleSinglePortInputOperator.java | 2 +-
.../apex/malhar/lib/io/SmtpOutputOperator.java | 2 +-
.../apex/malhar/lib/io/WebSocketInputOperator.java | 2 +-
.../malhar/lib/io/WebSocketOutputOperator.java | 2 +-
.../lib/io/WebSocketServerInputOperator.java | 2 +-
.../apex/malhar/lib/io/WidgetOutputOperator.java | 2 +-
.../malhar/lib/io/block/AbstractBlockReader.java | 6 +-
.../malhar/lib/io/block/AbstractFSBlockReader.java | 2 +-
.../apex/malhar/lib/io/block/BlockMetadata.java | 2 +-
.../apex/malhar/lib/io/block/BlockWriter.java | 6 +-
.../apex/malhar/lib/io/block/FSSliceReader.java | 2 +-
.../apex/malhar/lib/io/block/PartFileWriter.java | 4 +-
.../apex/malhar/lib/io/block/ReaderContext.java | 2 +-
.../apex/malhar/lib/io/block/package-info.java | 2 +-
.../lib/io/fs/AbstractFileInputOperator.java | 7 +-
.../lib/io/fs/AbstractFileOutputOperator.java | 4 +-
.../malhar/lib/io/fs/AbstractFileSplitter.java | 5 +-
.../apex/malhar/lib/io/fs/AbstractReconciler.java | 2 +-
.../io/fs/AbstractSingleFileOutputOperator.java | 2 +-
.../io/fs/AbstractThroughputFileInputOperator.java | 4 +-
.../io/fs/AbstractWindowFileOutputOperator.java | 2 +-
.../apex/malhar/lib/io/fs/FSInputModule.java | 13 +-
.../apache/apex/malhar/lib/io/fs/FileMerger.java | 4 +-
.../apache/apex/malhar/lib/io/fs/FileSplitter.java | 6 +-
.../apex/malhar/lib/io/fs/FileSplitterBase.java | 2 +-
.../apex/malhar/lib/io/fs/FileSplitterInput.java | 2 +-
.../apache/apex/malhar/lib/io/fs/FileStitcher.java | 8 +-
.../apex/malhar/lib/io/fs/FilterStreamCodec.java | 2 +-
.../apex/malhar/lib/io/fs/FilterStreamContext.java | 2 +-
.../malhar/lib/io/fs/FilterStreamProvider.java | 2 +-
.../apex/malhar/lib/io/fs/HDFSFileCopyModule.java | 10 +-
.../apex/malhar/lib/io/fs/HDFSFileMerger.java | 4 +-
.../apex/malhar/lib/io/fs/S3BlockReader.java | 8 +-
.../apex/malhar/lib/io/fs/S3InputModule.java | 4 +-
.../apache/apex/malhar/lib/io/fs/Synchronizer.java | 10 +-
.../apex/malhar/lib/io/fs/TailFsInputOperator.java | 2 +-
.../apache/apex/malhar/lib/io/fs/package-info.java | 2 +-
.../lib/io/jms/AbstractJMSInputOperator.java | 4 +-
.../lib/io/jms/AbstractJMSOutputOperator.java | 2 +-
.../jms/AbstractJMSSinglePortOutputOperator.java | 2 +-
.../lib/io/jms/FSPsuedoTransactionableStore.java | 2 +-
.../org/apache/apex/malhar/lib/io/jms/JMSBase.java | 2 +-
.../lib/io/jms/JMSBaseTransactionableStore.java | 4 +-
.../lib/io/jms/JMSMultiPortOutputOperator.java | 2 +-
.../malhar/lib/io/jms/JMSObjectInputOperator.java | 2 +-
.../malhar/lib/io/jms/JMSStringInputOperator.java | 2 +-
.../malhar/lib/io/jms/JMSTransactionableStore.java | 2 +-
.../apex/malhar/lib/io/jms/package-info.java | 2 +-
.../apache/apex/malhar/lib/io/package-info.java | 2 +-
.../apex/malhar/lib/join/AbstractJoinOperator.java | 2 +-
.../AbstractManagedStateInnerJoinOperator.java | 3 +-
.../apex/malhar/lib/join/AntiJoinOperator.java | 6 +-
.../org/apache/apex/malhar/lib/join/Bucket.java | 2 +-
.../apache/apex/malhar/lib/join/InMemoryStore.java | 2 +-
.../org/apache/apex/malhar/lib/join/JoinStore.java | 2 +-
.../apex/malhar/lib/join/JoinStreamCodec.java | 4 +-
.../apex/malhar/lib/join/MapJoinOperator.java | 2 +-
.../malhar/lib/join/POJOInnerJoinOperator.java | 2 +-
.../apex/malhar/lib/join/POJOJoinOperator.java | 5 +-
.../apex/malhar/lib/join/SemiJoinOperator.java | 6 +-
.../apex/malhar/lib/join/TimeBasedStore.java | 2 +-
.../org/apache/apex/malhar/lib/join/TimeEvent.java | 2 +-
.../apache/apex/malhar/lib/join/TimeEventImpl.java | 2 +-
.../lib/logs/ApacheLogParseMapOutputOperator.java | 2 +-
.../malhar/lib/logs/ApacheLogParseOperator.java | 2 +-
.../lib/logs/ApacheVirtualLogParseOperator.java | 4 +-
.../lib/logs/DimensionAggregationUnifier.java | 2 +-
.../apex/malhar/lib/logs/DimensionObject.java | 2 +-
.../lib/logs/FilteredLineToTokenArrayList.java | 2 +-
.../lib/logs/FilteredLineToTokenHashMap.java | 2 +-
.../lib/logs/FilteredLineTokenizerKeyVal.java | 2 +-
.../apex/malhar/lib/logs/InformationExtractor.java | 2 +-
.../apex/malhar/lib/logs/LineToTokenArrayList.java | 8 +-
.../apex/malhar/lib/logs/LineToTokenHashMap.java | 8 +-
.../apache/apex/malhar/lib/logs/LineTokenizer.java | 5 +-
.../apex/malhar/lib/logs/LineTokenizerKeyVal.java | 7 +-
.../lib/logs/MultiWindowDimensionAggregation.java | 4 +-
.../malhar/lib/logs/RegexMatchMapOperator.java | 2 +-
.../apache/apex/malhar/lib/logs/package-info.java | 2 +-
.../malhar/lib/math/AbstractAggregateCalc.java | 2 +-
.../apex/malhar/lib/math/AbstractOutput.java | 2 +-
.../lib/math/AbstractXmlCartesianProduct.java | 5 +-
.../math/AbstractXmlKeyValueCartesianProduct.java | 2 +-
.../org/apache/apex/malhar/lib/math/Average.java | 5 +-
.../apache/apex/malhar/lib/math/AverageKeyVal.java | 6 +-
.../org/apache/apex/malhar/lib/math/Division.java | 2 +-
.../apex/malhar/lib/math/LogicalCompare.java | 2 +-
.../malhar/lib/math/LogicalCompareToConstant.java | 2 +-
.../org/apache/apex/malhar/lib/math/Margin.java | 5 +-
.../apache/apex/malhar/lib/math/MarginKeyVal.java | 6 +-
.../org/apache/apex/malhar/lib/math/MarginMap.java | 6 +-
.../java/org/apache/apex/malhar/lib/math/Max.java | 5 +-
.../org/apache/apex/malhar/lib/math/MaxKeyVal.java | 7 +-
.../java/org/apache/apex/malhar/lib/math/Min.java | 5 +-
.../org/apache/apex/malhar/lib/math/MinKeyVal.java | 7 +-
.../apex/malhar/lib/math/MultiplyByConstant.java | 2 +-
.../org/apache/apex/malhar/lib/math/Range.java | 9 +-
.../apache/apex/malhar/lib/math/RangeKeyVal.java | 11 +-
.../apex/malhar/lib/math/RunningAverage.java | 2 +-
.../org/apache/apex/malhar/lib/math/Sigma.java | 2 +-
.../lib/math/SingleVariableAbstractCalculus.java | 2 +-
.../apex/malhar/lib/math/SquareCalculus.java | 2 +-
.../java/org/apache/apex/malhar/lib/math/Sum.java | 7 +-
.../org/apache/apex/malhar/lib/math/SumKeyVal.java | 6 +-
.../math/XmlKeyValueStringCartesianProduct.java | 2 +-
.../apache/apex/malhar/lib/math/package-info.java | 2 +-
.../apex/malhar/lib/metric/AvgAggregator.java | 2 +-
.../malhar/lib/metric/max/DoubleMaxAggregator.java | 2 +-
.../malhar/lib/metric/max/FloatMaxAggregator.java | 2 +-
.../malhar/lib/metric/max/IntMaxAggregator.java | 2 +-
.../malhar/lib/metric/max/LongMaxAggregator.java | 2 +-
.../malhar/lib/metric/min/DoubleMinAggregator.java | 2 +-
.../malhar/lib/metric/min/FloatMinAggregator.java | 2 +-
.../malhar/lib/metric/min/IntMinAggregator.java | 2 +-
.../malhar/lib/metric/min/LongMinAggregator.java | 2 +-
.../apex/malhar/lib/metric/package-info.java | 2 +-
.../lib/multiwindow/AbstractSlidingWindow.java | 2 +-
.../multiwindow/AbstractSlidingWindowKeyVal.java | 7 +-
.../lib/multiwindow/MultiWindowRangeKeyVal.java | 8 +-
.../lib/multiwindow/MultiWindowSumKeyVal.java | 7 +-
.../lib/multiwindow/SimpleMovingAverage.java | 5 +-
.../lib/multiwindow/SimpleMovingAverageObject.java | 2 +-
.../malhar/lib/multiwindow/SortedMovingWindow.java | 2 +-
.../apex/malhar/lib/multiwindow/package-info.java | 2 +-
.../org/apache/apex/malhar/lib/parser/Parser.java | 4 +-
.../apache/apex/malhar/lib/parser/XmlParser.java | 5 +-
.../apex/malhar/lib/parser/package-info.java | 2 +-
.../StatelessThroughputBasedPartitioner.java | 2 +-
.../StatsAwareStatelessPartitioner.java | 2 +-
.../apex/malhar/lib/partitioner/package-info.java | 2 +-
.../malhar/lib/projection/ProjectionOperator.java | 5 +-
.../apex/malhar/lib/script/JavaScriptOperator.java | 2 +-
.../apex/malhar/lib/script/ScriptOperator.java | 2 +-
.../apex/malhar/lib/script/package-info.java | 2 +-
.../state/managed/AbstractManagedStateImpl.java | 7 +-
.../apex/malhar/lib/state/managed/Bucket.java | 2 +-
.../lib/state/managed/BucketsFileSystem.java | 2 +-
.../lib/state/managed/ManagedStateContext.java | 3 +-
.../state/managed/ManagedTimeStateMultiValue.java | 2 +-
.../state/managed/ManagedTimeUnifiedStateImpl.java | 2 +-
.../apex/malhar/lib/statistics/MedianOperator.java | 2 +-
.../apex/malhar/lib/statistics/ModeOperator.java | 2 +-
.../malhar/lib/statistics/StandardDeviation.java | 2 +-
.../lib/statistics/WeightedMeanOperator.java | 6 +-
.../apex/malhar/lib/statistics/package-info.java | 2 +-
.../apex/malhar/lib/stream/AbstractAggregator.java | 2 +-
.../malhar/lib/stream/ArrayListAggregator.java | 2 +-
.../apex/malhar/lib/stream/ArrayListToItem.java | 5 +-
.../apex/malhar/lib/stream/ConsolidatorKeyVal.java | 5 +-
.../org/apache/apex/malhar/lib/stream/Counter.java | 2 +-
.../org/apache/apex/malhar/lib/stream/DevNull.java | 2 +-
.../apex/malhar/lib/stream/DevNullCounter.java | 2 +-
.../malhar/lib/stream/HashMapToKeyValPair.java | 7 +-
.../malhar/lib/stream/JsonByteArrayOperator.java | 2 +-
.../malhar/lib/stream/KeyValPairToHashMap.java | 7 +-
.../apex/malhar/lib/stream/RoundRobinHashMap.java | 5 +-
.../apex/malhar/lib/stream/StreamDuplicater.java | 5 +-
.../apex/malhar/lib/stream/StreamMerger.java | 2 +-
.../apex/malhar/lib/stream/package-info.java | 2 +-
.../lib/streamquery/condition/Condition.java | 2 +-
.../condition/JoinColumnEqualCondition.java | 2 +-
.../malhar/lib/streamquery/index/ColumnIndex.java | 2 +-
.../apex/malhar/lib/streamquery/index/Index.java | 2 +-
.../malhar/lib/testbench/ArrayListTestSink.java | 2 +-
.../malhar/lib/testbench/CollectorTestSink.java | 2 +-
.../malhar/lib/testbench/CompareFilterTuples.java | 2 +-
.../lib/testbench/CountAndLastTupleTestSink.java | 2 +-
.../apex/malhar/lib/testbench/CountOccurance.java | 2 +-
.../apex/malhar/lib/testbench/CountTestSink.java | 2 +-
.../apex/malhar/lib/testbench/EventClassifier.java | 2 +-
.../EventClassifierNumberToHashDouble.java | 2 +-
.../apex/malhar/lib/testbench/EventGenerator.java | 2 +-
.../malhar/lib/testbench/EventIncrementer.java | 5 +-
.../malhar/lib/testbench/FilterClassifier.java | 2 +-
.../lib/testbench/FilteredEventClassifier.java | 2 +-
.../apex/malhar/lib/testbench/HashTestSink.java | 2 +-
.../malhar/lib/testbench/HttpStatusFilter.java | 2 +-
.../apex/malhar/lib/testbench/KeyValSum.java | 2 +-
.../malhar/lib/testbench/RandomEventGenerator.java | 2 +-
.../malhar/lib/testbench/RandomWordGenerator.java | 2 +-
.../apex/malhar/lib/testbench/RedisSumOper.java | 2 +-
.../malhar/lib/testbench/SeedEventClassifier.java | 2 +-
.../malhar/lib/testbench/SeedEventGenerator.java | 5 +-
.../apex/malhar/lib/testbench/SumTestSink.java | 2 +-
.../malhar/lib/testbench/ThroughputCounter.java | 2 +-
.../apex/malhar/lib/testbench/TopOccurrence.java | 2 +-
.../apex/malhar/lib/testbench/package-info.java | 2 +-
.../malhar/lib/transform/TransformOperator.java | 6 +-
.../malhar/lib/util/AbstractBaseFrequentKey.java | 2 +-
.../lib/util/AbstractBaseFrequentKeyValueMap.java | 2 +-
.../malhar/lib/util/AbstractBaseMatchOperator.java | 2 +-
.../util/AbstractBaseNNonUniqueOperatorMap.java | 2 +-
.../malhar/lib/util/AbstractBaseNOperatorMap.java | 2 +-
.../lib/util/AbstractBaseNUniqueOperatorMap.java | 2 +-
.../malhar/lib/util/AbstractBaseSortOperator.java | 2 +-
.../util/AbstractDimensionTimeBucketOperator.java | 2 +-
.../lib/util/AbstractKeyValueStorageAgent.java | 2 +-
.../malhar/lib/util/ActiveMQMessageListener.java | 2 +-
.../lib/util/ActiveMQMultiTypeMessageListener.java | 2 +-
.../apex/malhar/lib/util/ArrayHashMapFrequent.java | 2 +-
.../lib/util/BaseFilteredKeyValueOperator.java | 2 +-
.../apex/malhar/lib/util/BaseKeyOperator.java | 2 +-
.../apex/malhar/lib/util/BaseKeyValueOperator.java | 5 +-
.../apex/malhar/lib/util/BaseLineTokenizer.java | 2 +-
.../apex/malhar/lib/util/BaseMatchOperator.java | 2 +-
.../lib/util/BaseNumberKeyValueOperator.java | 2 +-
.../malhar/lib/util/BaseNumberValueOperator.java | 2 +-
.../apex/malhar/lib/util/BaseUniqueKeyCounter.java | 2 +-
.../malhar/lib/util/BaseUniqueKeyValueCounter.java | 2 +-
.../lib/util/DimensionTimeBucketSumOperator.java | 2 +-
.../org/apache/apex/malhar/lib/util/FieldInfo.java | 2 +-
.../apex/malhar/lib/util/FieldValueGenerator.java | 6 +-
.../apex/malhar/lib/util/FilterOperator.java | 2 +-
.../org/apache/apex/malhar/lib/util/HighLow.java | 2 +-
.../malhar/lib/util/JavaScriptFilterOperator.java | 2 +-
.../apex/malhar/lib/util/KeyHashValPair.java | 2 +-
.../apache/apex/malhar/lib/util/KeyValPair.java | 2 +-
.../apex/malhar/lib/util/KryoCloneUtils.java | 2 +-
.../org/apache/apex/malhar/lib/util/PojoUtils.java | 9 +-
.../apex/malhar/lib/util/ReusableStringReader.java | 2 +-
.../apex/malhar/lib/util/ReversibleComparator.java | 2 +-
.../malhar/lib/util/StorageAgentKeyValueStore.java | 4 +-
.../org/apache/apex/malhar/lib/util/TableInfo.java | 2 +-
.../apache/apex/malhar/lib/util/TimeBucketKey.java | 2 +-
.../org/apache/apex/malhar/lib/util/TopNSort.java | 2 +-
.../apex/malhar/lib/util/TopNUniqueSort.java | 2 +-
.../lib/util/UnifierArrayHashMapFrequent.java | 2 +-
.../apex/malhar/lib/util/UnifierArrayList.java | 2 +-
.../apex/malhar/lib/util/UnifierBooleanAnd.java | 2 +-
.../apex/malhar/lib/util/UnifierBooleanOr.java | 2 +-
.../apex/malhar/lib/util/UnifierCountOccurKey.java | 2 +-
.../apex/malhar/lib/util/UnifierHashMap.java | 2 +-
.../malhar/lib/util/UnifierHashMapFrequent.java | 2 +-
.../malhar/lib/util/UnifierHashMapInteger.java | 2 +-
.../apex/malhar/lib/util/UnifierHashMapRange.java | 2 +-
.../malhar/lib/util/UnifierHashMapSumKeys.java | 2 +-
.../apex/malhar/lib/util/UnifierKeyValRange.java | 2 +-
.../apache/apex/malhar/lib/util/UnifierMap.java | 2 +-
.../apache/apex/malhar/lib/util/UnifierRange.java | 2 +-
.../apex/malhar/lib/util/UnifierSumNumber.java | 2 +-
.../lib/util/comparator/SliceComparator.java | 2 +-
.../apache/apex/malhar/lib/util/package-info.java | 2 +-
.../apex/malhar/lib/util/time/WindowUtils.java | 2 +-
.../apex/malhar/lib/wal/FSWindowDataManager.java | 4 +-
.../apex/malhar/lib/wal/WindowDataManager.java | 3 +-
.../lib/window/accumulation/AbstractPojoJoin.java | 5 +-
.../lib/window/accumulation/PojoFullOuterJoin.java | 5 +-
.../lib/window/accumulation/PojoInnerJoin.java | 6 +-
.../lib/window/accumulation/PojoLeftOuterJoin.java | 5 +-
.../window/accumulation/PojoRightOuterJoin.java | 5 +-
.../malhar/lib/window/accumulation/TopNByKey.java | 2 +-
.../impl/KeyedWindowedMergeOperatorImpl.java | 2 +-
.../lib/window/impl/KeyedWindowedOperatorImpl.java | 2 +-
.../window/impl/WindowedMergeOperatorFeatures.java | 3 +-
.../malhar/lib/xml/AbstractXmlDOMOperator.java | 2 +-
.../org/apache/hadoop/io/file/tfile/DTFile.java | 7 +-
.../apex/malhar/lib/algo/BottomNMapTest.java | 6 +-
.../apex/malhar/lib/algo/BottomNUnifierTest.java | 4 +-
.../apex/malhar/lib/algo/BottomNUniqueMapTest.java | 6 +-
.../malhar/lib/algo/CompareExceptCountMapTest.java | 6 +-
.../apache/apex/malhar/lib/algo/DistinctTest.java | 8 +-
.../apex/malhar/lib/algo/FilterValuesTest.java | 6 +-
.../apex/malhar/lib/algo/InsertSortTest.java | 6 +-
.../apex/malhar/lib/algo/LeastFrequentKeyTest.java | 6 +-
.../apex/malhar/lib/algo/MatchAllMapTest.java | 6 +-
.../apex/malhar/lib/algo/MatchAnyMapTest.java | 6 +-
.../apex/malhar/lib/algo/MergeSortNumberTest.java | 6 +-
.../apex/malhar/lib/algo/MostFrequentKeyTest.java | 6 +-
.../org/apache/apex/malhar/lib/algo/TopNTest.java | 6 +-
.../apex/malhar/lib/algo/TopNUniqueTest.java | 6 +-
.../apex/malhar/lib/algo/UniqueCounterTest.java | 6 +-
.../apex/malhar/lib/algo/UniqueValueCountTest.java | 12 +-
.../apex/malhar/lib/appdata/ThreadUtils.java | 2 +-
.../datastructs/CacheLRUSynchronousFlushTest.java | 6 +-
.../appdata/datastructs/DimensionalTableTest.java | 6 +-
.../dimensions/CustomTimeBucketRegistryTest.java | 7 +-
.../appdata/dimensions/DimensionsEventTest.java | 9 +-
.../malhar/lib/appdata/gpo/GPOMutableTest.java | 8 +-
.../apex/malhar/lib/appdata/gpo/GPOUtilsTest.java | 9 +-
.../lib/appdata/gpo/SerdeFieldsDescriptorTest.java | 7 +-
.../lib/appdata/gpo/SerdeListGPOMutableTest.java | 7 +-
.../lib/appdata/gpo/SerdeListPrimitiveTest.java | 2 +-
.../lib/appdata/gpo/SerdeListStringTest.java | 2 +-
.../apex/malhar/lib/appdata/query/MockQuery.java | 4 +-
.../apex/malhar/lib/appdata/query/MockResult.java | 10 +-
.../lib/appdata/query/MockResultSerializer.java | 10 +-
.../query/QueryManagerAsynchronousTest.java | 10 +-
.../appdata/query/QueryManagerSynchronousTest.java | 8 +-
.../malhar/lib/appdata/query/QueueUtilsTest.java | 6 +-
.../query/SimpleDoneQueryQueueManagerTest.java | 7 +-
.../appdata/query/SimpleQueryQueueManagerTest.java | 4 +-
.../lib/appdata/query/WEQueryQueueManagerTest.java | 5 +-
.../appdata/query/WindowBoundedServiceTest.java | 4 +-
.../serde/MessageDeserializerFactoryTest.java | 6 +-
.../query/serde/SimpleDataValidatorTest.java | 4 +-
.../lib/appdata/schemas/CustomTimeBucketTest.java | 2 +-
.../schemas/DataQuerySnapshotDeserializerTest.java | 8 +-
.../lib/appdata/schemas/DataQuerySnapshotTest.java | 2 +-
.../schemas/DataResultSnapshotSerializerTest.java | 8 +-
.../DimensionalConfigurationSchemaTest.java | 9 +-
.../lib/appdata/schemas/DimensionalSchemaTest.java | 5 +-
.../lib/appdata/schemas/FieldsDescriptorTest.java | 2 +-
.../lib/appdata/schemas/ResultFormatterTest.java | 4 +-
.../lib/appdata/schemas/SchemaQueryTest.java | 6 +-
.../schemas/SchemaRegistryMultipleTest.java | 6 +-
.../appdata/schemas/SchemaRegistrySingleTest.java | 4 +-
.../lib/appdata/schemas/SchemaTestUtils.java | 2 +-
.../lib/appdata/schemas/SchemaUtilsTest.java | 2 +-
.../lib/appdata/schemas/SnapshotSchemaTest.java | 2 +-
.../malhar/lib/appdata/schemas/TimeBucketTest.java | 4 +-
.../snapshot/AppDataSnapshotServerMapTest.java | 8 +-
.../snapshot/AppDataSnapshotServerPojoTest.java | 14 +-
.../AppDataSnapshotServerTagsSupportTest.java | 12 +-
.../malhar/lib/bandwidth/BandwidthManagerTest.java | 5 +-
.../lib/bandwidth/BandwidthPartitionerTest.java | 2 +-
.../codec/JavaSerializationStreamCodecTest.java | 4 +-
.../apex/malhar/lib/codec/KryoStreamCodecTest.java | 2 +-
.../converter/ByteArrayToStringConverterTest.java | 6 +-
.../converter/MapToKeyValuePairConverterTest.java | 8 +-
.../StringValueToNumberConverterForMapTest.java | 6 +-
.../malhar/lib/counters/BasicCountersTest.java | 2 +-
.../malhar/lib/db/KeyValueStoreOperatorTest.java | 4 +-
.../TransactionableKeyValueStoreOperatorTest.java | 4 +-
.../apex/malhar/lib/db/cache/CacheManagerTest.java | 2 +-
.../apex/malhar/lib/db/cache/CacheStoreTest.java | 3 +-
.../db/jdbc/JDBCLookupCacheBackedOperatorTest.java | 7 +-
.../apache/apex/malhar/lib/db/jdbc/JdbcIOApp.java | 6 +-
.../apex/malhar/lib/db/jdbc/JdbcIOAppTest.java | 2 +-
.../db/jdbc/JdbcInputOperatorApplicationTest.java | 4 +-
...dbcNonTransactionalBatchOutputOperatorTest.java | 17 +-
.../JdbcNonTransactionalOutputOperatorTest.java | 4 +-
.../lib/db/jdbc/JdbcNonTransactionalStoreTest.java | 2 +-
.../apex/malhar/lib/db/jdbc/JdbcOperatorTest.java | 2 +-
.../db/jdbc/JdbcPojoOperatorApplicationTest.java | 2 +-
.../malhar/lib/db/jdbc/JdbcPojoOperatorTest.java | 16 +-
.../lib/db/jdbc/JdbcPojoPollableOpeartorTest.java | 14 +-
.../apex/malhar/lib/db/jdbc/JdbcStoreTest.java | 2 +-
.../RedshiftJdbcTransactionalOperatorTest.java | 4 +-
.../lib/dedup/DeduperBoundedPOJOImplTest.java | 8 +-
.../apex/malhar/lib/dedup/DeduperOrderingTest.java | 4 +-
.../malhar/lib/dedup/DeduperPartitioningTest.java | 2 +-
.../lib/dedup/DeduperTimeBasedPOJOImplTest.java | 12 +-
.../dimensions/CustomTimeBucketRegistryTest.java | 4 +-
.../lib/dimensions/DimensionsDescriptorTest.java | 10 +-
.../apex/malhar/lib/filter/FilterAppTest.java | 5 +-
.../apache/apex/malhar/lib/filter/FilterTest.java | 5 +-
.../malhar/lib/formatter/JsonFormatterTest.java | 13 +-
.../malhar/lib/formatter/XmlFormatterTest.java | 8 +-
.../lib/fs/FSRecordCompactionOperatorTest.java | 4 +-
.../lib/fs/GenericFileOutputOperatorTest.java | 2 +-
.../lib/fs/s3/S3DelimitedRecordReaderTest.java | 9 +-
.../lib/fs/s3/S3FixedWidthRecordReaderTest.java | 9 +-
.../fs/s3/S3InitiateFileUploadOperatorTest.java | 8 +-
.../malhar/lib/fs/s3/S3OutputModuleMockTest.java | 4 +-
.../apex/malhar/lib/fs/s3/S3ReconcilerTest.java | 4 +-
.../malhar/lib/fs/s3/S3RecordReaderMockTest.java | 6 +-
.../lib/helper/OperatorContextTestHelper.java | 2 +-
.../lib/helper/SamplePubSubWebSocketServlet.java | 2 +-
.../apex/malhar/lib/helper/TestPortContext.java | 2 +-
.../apex/malhar/lib/io/ApacheRandomLogsTest.java | 4 +-
.../malhar/lib/io/FTPStringInputOperatorTest.java | 10 +-
.../apex/malhar/lib/io/HttpGetMapOperatorTest.java | 6 +-
.../lib/io/HttpJsonChunksInputOperatorTest.java | 5 +-
.../malhar/lib/io/HttpLinesInputOperatorTest.java | 7 +-
.../lib/io/HttpMultiValuedMapGetOperatorTest.java | 8 +-
.../malhar/lib/io/HttpPostOutputOperatorTest.java | 4 +-
.../lib/io/IdempotentStorageManagerTest.java | 4 +-
.../lib/io/PubSubWebSocketAppDataOperatorTest.java | 2 +-
.../lib/io/PubSubWebSocketAppDataQueryTest.java | 4 +-
.../lib/io/PubSubWebSocketAppDataResultTest.java | 2 +-
.../malhar/lib/io/PubSubWebSocketOperatorTest.java | 9 +-
.../apex/malhar/lib/io/SmtpOutputOperatorTest.java | 2 +-
.../malhar/lib/io/SocketInputOperatorTest.java | 6 +-
.../lib/io/WebSocketServerInputOperatorTest.java | 2 +-
.../lib/io/block/AbstractBlockReaderTest.java | 6 +-
.../apex/malhar/lib/io/block/FSLineReaderTest.java | 7 +-
.../malhar/lib/io/block/FSSliceReaderTest.java | 6 +-
.../malhar/lib/io/block/PartFileWriterTest.java | 5 +-
.../lib/io/block/ReadAheadLineReaderTest.java | 2 +-
...stractFileInputOperatorFailureHandlingTest.java | 10 +-
.../lib/io/fs/AbstractFileInputOperatorTest.java | 15 +-
.../lib/io/fs/AbstractFileOutputOperatorTest.java | 10 +-
.../malhar/lib/io/fs/AbstractReconcilerTest.java | 6 +-
.../fs/AbstractSingleFileOutputOperatorTest.java | 8 +-
.../AbstractThroughputFileInputOperatorTest.java | 8 +-
.../fs/AbstractWindowFileOutputOperatorTest.java | 9 +-
.../apex/malhar/lib/io/fs/BlockWriterTest.java | 8 +-
.../malhar/lib/io/fs/FSInputModuleAppTest.java | 11 +-
.../apex/malhar/lib/io/fs/FSInputModuleTest.java | 6 +-
.../lib/io/fs/FastMergerDecisionMakerTest.java | 9 +-
.../apex/malhar/lib/io/fs/FileMergerTest.java | 18 +-
.../malhar/lib/io/fs/FileSplitterBaseTest.java | 10 +-
.../malhar/lib/io/fs/FileSplitterInputTest.java | 12 +-
.../apex/malhar/lib/io/fs/FileSplitterTest.java | 10 +-
.../apex/malhar/lib/io/fs/FileStitcherTest.java | 18 +-
.../malhar/lib/io/fs/S3InputModuleAppTest.java | 8 +-
.../apex/malhar/lib/io/fs/SynchronizerTest.java | 12 +-
.../malhar/lib/io/fs/TailFsInputOperatorTest.java | 6 +-
.../io/jms/FSPsuedoTransactionableStoreTest.java | 2 +-
.../lib/io/jms/JMSMultiPortOutputOperatorTest.java | 9 +-
.../lib/io/jms/JMSObjectInputOperatorTest.java | 7 +-
.../malhar/lib/io/jms/JMSOutputOperatorTest.java | 6 +-
.../lib/io/jms/JMSStringInputOperatorTest.java | 6 +-
.../apache/apex/malhar/lib/io/jms/JMSTestBase.java | 2 +-
.../lib/io/jms/JMSTransactionableStoreTest.java | 2 +-
.../io/jms/JMSTransactionableStoreTestBase.java | 11 +-
.../lib/io/jms/SQSStringInputOperatorTest.java | 10 +-
.../apache/apex/malhar/lib/io/jms/SQSTestBase.java | 2 +-
.../apex/malhar/lib/join/AntiJoinOperatorTest.java | 12 +-
.../malhar/lib/join/MapTimeBasedJoinOperator.java | 9 +-
.../malhar/lib/join/POJOInnerJoinOperatorTest.java | 8 +-
.../lib/join/POJOPartitionJoinOperatorTest.java | 2 +-
.../lib/join/POJOTimeBasedJoinOperatorTest.java | 8 +-
.../apex/malhar/lib/join/SemiJoinOperatorTest.java | 12 +-
.../logs/ApacheLogParseMapOutputOperatorTest.java | 6 +-
.../lib/logs/ApacheLogParseOperatorTest.java | 6 +-
.../lib/logs/FilteredLineToTokenArrayListTest.java | 6 +-
.../lib/logs/FilteredLineToTokenHashMapTest.java | 6 +-
.../lib/logs/FilteredLineTokenizerKeyValTest.java | 6 +-
.../malhar/lib/logs/LineToTokenArrayListTest.java | 6 +-
.../malhar/lib/logs/LineToTokenHashMapTest.java | 6 +-
.../malhar/lib/logs/LineTokenizerKeyValTest.java | 6 +-
.../apex/malhar/lib/logs/LineTokenizerTest.java | 6 +-
.../logs/MultiWindowDimensionAggregationTest.java | 7 +-
.../malhar/lib/logs/RegexMatchMapOperatorTest.java | 6 +-
.../malhar/lib/logs/TopNUniqueSiteStatsTest.java | 7 +-
.../apex/malhar/lib/math/AverageKeyValTest.java | 8 +-
.../apache/apex/malhar/lib/math/AverageTest.java | 8 +-
.../apex/malhar/lib/math/CountOccuranceTest.java | 6 +-
.../apache/apex/malhar/lib/math/DivisionTest.java | 6 +-
.../apex/malhar/lib/math/LogicalCompareTest.java | 7 +-
.../lib/math/LogicalCompareToConstantTest.java | 6 +-
.../apex/malhar/lib/math/MarginKeyValTest.java | 8 +-
.../apache/apex/malhar/lib/math/MarginMapTest.java | 6 +-
.../apache/apex/malhar/lib/math/MarginTest.java | 6 +-
.../apache/apex/malhar/lib/math/MaxKeyValTest.java | 9 +-
.../org/apache/apex/malhar/lib/math/MaxTest.java | 6 +-
.../apache/apex/malhar/lib/math/MinKeyValTest.java | 9 +-
.../org/apache/apex/malhar/lib/math/MinTest.java | 6 +-
.../malhar/lib/math/MultiplyByConstantTest.java | 6 +-
.../apex/malhar/lib/math/RangeKeyValTest.java | 10 +-
.../org/apache/apex/malhar/lib/math/RangeTest.java | 8 +-
.../apex/malhar/lib/math/RunningAverageTest.java | 4 +-
.../org/apache/apex/malhar/lib/math/SigmaTest.java | 6 +-
.../apex/malhar/lib/math/SquareCalculusTest.java | 6 +-
.../apache/apex/malhar/lib/math/SumKeyValTest.java | 8 +-
.../org/apache/apex/malhar/lib/math/SumTest.java | 6 +-
.../XmlKeyValueStringCartesianProductTest.java | 2 +-
.../multiwindow/MultiWindowRangeKeyValTest.java | 8 +-
.../lib/multiwindow/MultiWindowSumKeyValTest.java | 8 +-
.../lib/multiwindow/SimpleMovingAverageTest.java | 8 +-
.../malhar/lib/multiwindow/SlidingWindowTest.java | 8 +-
.../lib/multiwindow/SortedMovingWindowTest.java | 5 +-
.../lib/parser/XmlParserApplicationTest.java | 12 +-
.../apex/malhar/lib/parser/XmlParserTest.java | 8 +-
.../lib/partitioner/StatelessPartitionerTest.java | 5 +-
.../StatelessThroughputBasedPartitionerTest.java | 5 +-
.../apex/malhar/lib/projection/ActivateTest.java | 2 +-
.../apex/malhar/lib/projection/ProjectionTest.java | 4 +-
.../malhar/lib/script/JavaScriptOperatorTest.java | 8 +-
.../lib/state/managed/BucketsFileSystemTest.java | 5 +-
.../lib/state/managed/DefaultBucketTest.java | 6 +-
.../managed/IncrementalCheckpointManagerTest.java | 6 +-
.../lib/state/managed/ManagedStateImplTest.java | 7 +-
.../lib/state/managed/ManagedStateTestUtils.java | 6 +-
.../state/managed/ManagedTimeStateImplTest.java | 8 +-
.../managed/ManagedTimeUnifiedStateImplTest.java | 4 +-
.../lib/state/managed/MockManagedStateContext.java | 7 +-
.../MovingBoundaryTimeBucketAssignerTest.java | 3 +-
.../malhar/lib/state/managed/StateTrackerTest.java | 5 +-
...SequentialSpillableIdentifierGeneratorTest.java | 2 +-
.../spillable/SpillableArrayListImplTest.java | 7 +-
.../SpillableArrayListMultimapImplTest.java | 4 +-
.../lib/state/spillable/SpillableMapImplTest.java | 4 +-
.../spillable/SpillableSetMultimapImplTest.java | 7 +-
.../lib/state/spillable/SpillableTestUtils.java | 4 +-
.../inmem/InMemorySpillableStateStoreTest.java | 2 +-
.../lib/statistics/MeridianOperatorTest.java | 8 +-
.../malhar/lib/statistics/ModeOperatorTest.java | 8 +-
.../statistics/StandardDeviationOperatorTest.java | 6 +-
.../lib/statistics/WeightedMeanOperatorTest.java | 8 +-
.../malhar/lib/stream/ArrayListAggregatorTest.java | 6 +-
.../malhar/lib/stream/ArrayListToItemTest.java | 6 +-
.../malhar/lib/stream/ConsolidatorKeyValTest.java | 8 +-
.../apache/apex/malhar/lib/stream/CounterTest.java | 6 +-
.../apex/malhar/lib/stream/DevNullCounterTest.java | 6 +-
.../apache/apex/malhar/lib/stream/DevNullTest.java | 6 +-
.../malhar/lib/stream/HashMapToKeyValPairTest.java | 6 +-
.../lib/stream/JsonByteArrayOperatorTest.java | 4 +-
.../malhar/lib/stream/KeyPairToHashMapTest.java | 8 +-
.../malhar/lib/stream/RoundRobinHashMapTest.java | 6 +-
.../malhar/lib/stream/StreamDuplicaterTest.java | 6 +-
.../apex/malhar/lib/stream/StreamMergerTest.java | 6 +-
.../lib/testbench/ActiveMQMessageGenerator.java | 2 +-
.../malhar/lib/testbench/EventClassifierTest.java | 4 +-
.../malhar/lib/testbench/EventGeneratorTest.java | 4 +-
.../malhar/lib/testbench/EventIncrementerTest.java | 4 +-
.../lib/testbench/FilteredEventClassifierTest.java | 4 +-
.../lib/testbench/RandomEventGeneratorTest.java | 4 +-
.../lib/testbench/RandomWordGeneratorTest.java | 2 +-
.../lib/testbench/SeedEventClassifierTest.java | 4 +-
.../lib/testbench/SeedEventGeneratorTest.java | 4 +-
.../lib/testbench/ThroughputCounterTest.java | 4 +-
.../lib/transform/TransformOperatorAppTest.java | 5 +-
.../lib/transform/TransformOperatorTest.java | 7 +-
.../util/DimensionTimeBucketSumOperatorTest.java | 5 +-
.../lib/util/JavaScriptFilterOperatorTest.java | 6 +-
.../apex/malhar/lib/util/KryoCloneUtilsTest.java | 2 +-
.../apache/apex/malhar/lib/util/PojoUtilsTest.java | 78 +++---
.../apex/malhar/lib/util/TestObjAllTypes.java | 2 +-
.../org/apache/apex/malhar/lib/util/TestUtils.java | 2 +-
.../apex/malhar/lib/util/time/WindowUtilsTest.java | 4 +-
.../malhar/lib/wal/FSWindowDataManagerTest.java | 5 +-
.../apex/malhar/lib/wal/FileSystemWALTest.java | 2 +-
.../lib/window/SpillableWindowedStorageTest.java | 2 +-
.../malhar/lib/window/WindowedOperatorTest.java | 4 +-
.../lib/window/accumulation/PojoInnerJoinTest.java | 4 +-
.../lib/window/accumulation/PojoOuterJoinTest.java | 4 +-
.../lib/window/accumulation/TopNByKeyTest.java | 2 +-
.../KeyedWindowedMergeOperatorTestApplication.java | 4 +-
.../lib/window/impl/WindowedMergeOperatorTest.java | 5 +-
.../impl/WindowedMergeOperatorTestApplication.java | 2 +-
.../malhar/lib/window/sample/pi/Application.java | 2 +-
.../lib/window/sample/wordcount/Application.java | 4 +-
pom.xml | 4 +-
samples/pom.xml | 55 -----
.../samples/lib/algo/AllAfterMatchMapSample.java | 65 -----
.../samples/lib/algo/BottomNMapSample.java | 65 -----
.../datatorrent/samples/lib/algo/package-info.java | 25 --
.../samples/lib/io/HttpInputOperatorSample.java | 64 -----
.../datatorrent/samples/lib/io/package-info.java | 24 --
.../samples/lib/math/AverageKeyValSample.java | 64 -----
.../samples/lib/math/AverageSample.java | 70 ------
.../datatorrent/samples/lib/math/ChangeSample.java | 70 ------
.../samples/lib/math/CompreMapSample.java | 61 -----
.../samples/lib/math/CountKeyValSample.java | 62 -----
.../samples/lib/math/CountKeyValues.java | 88 -------
.../samples/lib/math/LogicalCompareSample.java | 62 -----
.../samples/lib/math/PartitionMathSumSample.java | 71 ------
.../lib/math/RandomEvenetGeneratorSample.java | 57 -----
.../samples/lib/math/RandomKeyValMap.java | 84 -------
.../samples/lib/math/RandomKeyValues.java | 84 -------
.../samples/lib/math/RandomPairIntegers.java | 84 -------
.../datatorrent/samples/lib/math/ScriptSample.java | 61 -----
.../samples/lib/math/SingleKeyValMap.java | 82 -------
.../datatorrent/samples/lib/math/package-info.java | 25 --
.../java/com/datatorrent/samples/package-info.java | 25 --
.../lib/math/PartitionMathSumSampleTest.java | 44 ----
samples/src/test/resources/log4j.properties | 43 ----
sql/pom.xml | 2 +-
.../malhar/sql/codegen/ExpressionCompiler.java | 2 +-
.../sql/operators/FilterTransformOperator.java | 6 +-
.../apex/malhar/sql/operators/LineReader.java | 2 +-
.../apex/malhar/sql/planner/ApexRelNode.java | 2 +-
.../apex/malhar/sql/table/CSVMessageFormat.java | 6 +-
.../apache/apex/malhar/sql/KafkaEndpointTest.java | 4 +-
.../java/org/apache/apex/malhar/sql/SerDeTest.java | 4 +-
.../apache/apex/malhar/sql/StreamEndpointTest.java | 4 +-
stream/pom.xml | 2 +-
.../apex/malhar/stream/api/WindowedStream.java | 3 +-
.../malhar/stream/api/impl/ApexStreamImpl.java | 2 +-
.../stream/api/impl/ApexWindowedStreamImpl.java | 3 +-
.../apex/malhar/stream/api/impl/StreamFactory.java | 2 +-
.../stream/sample/ApplicationWithStreamAPI.java | 2 +-
.../sample/LocalTestWithoutStreamApplication.java | 3 +-
.../apex/malhar/stream/sample/MyStreamTest.java | 3 +-
.../stream/sample/WordCountWithStreamAPI.java | 2 +-
1576 files changed, 6492 insertions(+), 7802 deletions(-)
diff --git a/apps/filecopy/pom.xml b/apps/filecopy/pom.xml
index d6a0ccd..69b2e95 100644
--- a/apps/filecopy/pom.xml
+++ b/apps/filecopy/pom.xml
@@ -24,7 +24,7 @@
<parent>
<artifactId>malhar-apps</artifactId>
<groupId>org.apache.apex</groupId>
- <version>3.9.0-SNAPSHOT</version>
+ <version>4.0.0-SNAPSHOT</version>
</parent>
<artifactId>filecopy</artifactId>
@@ -79,7 +79,7 @@
<dependency>
<groupId>org.apache.apex</groupId>
<artifactId>malhar-library</artifactId>
- <version>3.9.0-SNAPSHOT</version>
+ <version>4.0.0-SNAPSHOT</version>
<exclusions>
<exclusion>
<groupId>*</groupId>
diff --git a/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java b/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java
index 2688f1c..bb1e1d9 100644
--- a/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java
+++ b/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java
@@ -19,14 +19,14 @@
package com.datatorrent.apps.copy;
+import org.apache.apex.malhar.lib.io.fs.FSInputModule;
+import org.apache.apex.malhar.lib.io.fs.HDFSFileCopyModule;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.io.fs.FSInputModule;
-import com.datatorrent.lib.io.fs.HDFSFileCopyModule;
/**
* Application for HDFS to HDFS file copy
diff --git a/apps/logstream/pom.xml b/apps/logstream/pom.xml
index 07a5eee..13cc865 100644
--- a/apps/logstream/pom.xml
+++ b/apps/logstream/pom.xml
@@ -24,7 +24,7 @@
<parent>
<artifactId>malhar-apps</artifactId>
<groupId>org.apache.apex</groupId>
- <version>3.9.0-SNAPSHOT</version>
+ <version>4.0.0-SNAPSHOT</version>
</parent>
<groupId>org.apache.apex</groupId>
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java
index 495987e..74997f4 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java
@@ -22,8 +22,8 @@ import java.util.ArrayList;
import java.util.HashMap;
import java.util.Map;
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.util.KeyValPair;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.util.KeyValPair;
import com.datatorrent.common.util.BaseOperator;
import com.datatorrent.api.DefaultInputPort;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java
index 82c9214..e3858b4 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java
@@ -28,25 +28,26 @@ import java.util.Set;
import org.apache.apex.malhar.contrib.misc.streamquery.SelectOperator;
import org.apache.apex.malhar.contrib.misc.streamquery.condition.EqualValueCondition;
import org.apache.apex.malhar.lib.utils.PubSubHelper;
+import org.apache.apex.malhar.contrib.redis.RedisKeyValPairOutputOperator;
+import org.apache.apex.malhar.contrib.redis.RedisMapOutputOperator;
+import org.apache.apex.malhar.contrib.redis.RedisNumberSummationMapOutputOperator;
+import org.apache.apex.malhar.lib.algo.TopN;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.io.PubSubWebSocketOutputOperator;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.logs.MultiWindowDimensionAggregation;
+import org.apache.apex.malhar.lib.logs.MultiWindowDimensionAggregation.AggregateOperation;
+import org.apache.apex.malhar.lib.stream.Counter;
+import org.apache.apex.malhar.lib.stream.JsonByteArrayOperator;
+import org.apache.apex.malhar.lib.streamquery.index.ColumnIndex;
+import org.apache.apex.malhar.lib.util.AbstractDimensionTimeBucketOperator;
+import org.apache.apex.malhar.lib.util.DimensionTimeBucketSumOperator;
+import org.apache.commons.lang.StringUtils;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.Operator.InputPort;
import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.contrib.redis.RedisKeyValPairOutputOperator;
-import com.datatorrent.contrib.redis.RedisMapOutputOperator;
-import com.datatorrent.contrib.redis.RedisNumberSummationMapOutputOperator;
-import com.datatorrent.lib.algo.TopN;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.io.PubSubWebSocketOutputOperator;
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.logs.MultiWindowDimensionAggregation;
-import com.datatorrent.lib.logs.MultiWindowDimensionAggregation.AggregateOperation;
-import com.datatorrent.lib.stream.Counter;
-import com.datatorrent.lib.stream.JsonByteArrayOperator;
-import com.datatorrent.lib.streamquery.index.ColumnIndex;
-import com.datatorrent.lib.util.AbstractDimensionTimeBucketOperator;
-import com.datatorrent.lib.util.DimensionTimeBucketSumOperator;
/**
* Log stream processing application based on Apex platform.<br>
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java
index 333b877..a458b3c 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java
@@ -18,11 +18,10 @@
*/
package com.datatorrent.apps.logstream;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.stream.JsonByteArrayOperator;
import org.apache.hadoop.conf.Configuration;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.stream.JsonByteArrayOperator;
-
import com.datatorrent.api.Context.PortContext;
import com.datatorrent.api.DAG;
import com.datatorrent.api.StreamingApplication;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java
index 2650795..de96781 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java
@@ -28,10 +28,11 @@ import com.google.common.collect.Sets;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.codec.KryoSerializableStreamCodec;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
import org.apache.commons.lang.mutable.MutableDouble;
-import com.datatorrent.lib.codec.KryoSerializableStreamCodec;
-import com.datatorrent.lib.logs.DimensionObject;
import com.datatorrent.api.*;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.Partitioner.PartitioningContext;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java
index e495ee6..606edb7 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java
@@ -26,10 +26,9 @@ import java.util.Map.Entry;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
import org.apache.commons.lang.mutable.MutableDouble;
-import com.datatorrent.lib.logs.DimensionObject;
-
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.Operator.Unifier;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java
index 57c558b..74dc3a6 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java
@@ -20,10 +20,12 @@ package com.datatorrent.apps.logstream;
import com.datatorrent.common.util.BaseOperator;
import com.datatorrent.api.DefaultInputPort;
-import com.datatorrent.lib.logs.DimensionObject;
+
import java.util.HashMap;
import java.util.Map;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+
/**
* Log Score
*
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java
index a09e64f..8df62fa 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java
@@ -28,9 +28,9 @@ import com.google.common.collect.Sets;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.datatorrent.lib.algo.TopN;
-import com.datatorrent.lib.codec.KryoSerializableStreamCodec;
-import com.datatorrent.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.algo.TopN;
+import org.apache.apex.malhar.lib.codec.KryoSerializableStreamCodec;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultPartition;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java
index 29c92e6..715cb6b 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java
@@ -26,11 +26,10 @@ import java.util.Map.Entry;
import javax.validation.constraints.NotNull;
+import org.apache.apex.malhar.lib.io.WidgetOutputOperator;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
import org.apache.commons.lang3.tuple.MutablePair;
-import com.datatorrent.lib.io.WidgetOutputOperator;
-import com.datatorrent.lib.logs.DimensionObject;
-
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultInputPort;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java
index a72dd05..a4e7fd2 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java
@@ -24,13 +24,14 @@ import org.codehaus.jettison.json.JSONObject;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.contrib.rabbitmq.AbstractSinglePortRabbitMQInputOperator;
+
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.Partitioner.PartitioningContext;
import com.datatorrent.api.DefaultPartition;
import com.datatorrent.api.Partitioner;
import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.contrib.rabbitmq.AbstractSinglePortRabbitMQInputOperator;
/**
*
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java
index 1ecd094..6eadc6e 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java
@@ -26,11 +26,10 @@ import java.util.Map.Entry;
import org.junit.Assert;
import org.junit.Test;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
import org.apache.commons.lang.mutable.MutableDouble;
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.testbench.CollectorTestSink;
-
import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
import com.datatorrent.netlet.util.DTThrowable;
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java
index a835dab..05929e2 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java
@@ -26,11 +26,10 @@ import org.junit.Assert;
import org.junit.Test;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
import org.apache.commons.lang.mutable.MutableDouble;
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.testbench.CollectorTestSink;
-
/**
*
* Tests logstream dimension operator unifier.
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java
index 2b7a5bd..a2bc229 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java
@@ -23,7 +23,7 @@ import org.junit.Assert;
import org.junit.Test;
-import com.datatorrent.lib.testbench.CollectorTestSink;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java
index 69d0567..d83373c 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java
@@ -26,11 +26,10 @@ import java.util.Map;
import org.junit.Assert;
import org.junit.Test;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
import org.apache.commons.lang.mutable.MutableDouble;
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.testbench.CollectorTestSink;
-
import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
/**
diff --git a/apps/pom.xml b/apps/pom.xml
index 679d623..1f46ecd 100644
--- a/apps/pom.xml
+++ b/apps/pom.xml
@@ -25,7 +25,7 @@
<parent>
<groupId>org.apache.apex</groupId>
<artifactId>malhar</artifactId>
- <version>3.9.0-SNAPSHOT</version>
+ <version>4.0.0-SNAPSHOT</version>
</parent>
<artifactId>malhar-apps</artifactId>
diff --git a/benchmark/pom.xml b/benchmark/pom.xml
index 0822549..1b921d7 100644
--- a/benchmark/pom.xml
+++ b/benchmark/pom.xml
@@ -24,7 +24,7 @@
<parent>
<artifactId>malhar</artifactId>
<groupId>org.apache.apex</groupId>
- <version>3.9.0-SNAPSHOT</version>
+ <version>4.0.0-SNAPSHOT</version>
</parent>
<groupId>org.apache.apex</groupId>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java b/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java
index aa10eea..1491f47 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import org.apache.hadoop.conf.Configuration;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java
index d8d51b8..6992bd8 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import org.apache.hadoop.conf.Configuration;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java
index bf5b876..6e29652 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import org.apache.hadoop.conf.Configuration;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java
index 4f12791..95eaacc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java
@@ -16,8 +16,9 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
@@ -25,7 +26,6 @@ import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
/**
*
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java
index 8ae0a94..12e43e2 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java
@@ -16,15 +16,15 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import java.util.ArrayList;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.datatorrent.contrib.couchbase.AbstractCouchBaseInputOperator;
-import com.datatorrent.contrib.couchbase.CouchBaseWindowStore;
+import org.apache.apex.malhar.contrib.couchbase.AbstractCouchBaseInputOperator;
+import org.apache.apex.malhar.contrib.couchbase.CouchBaseWindowStore;
/**
* <p>CouchBaseInputOperator class.</p>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java
index 8b6fc49..43a8e91 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java
@@ -16,9 +16,9 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
-import com.datatorrent.contrib.couchbase.AbstractCouchBaseSetOperator;
+import org.apache.apex.malhar.contrib.couchbase.AbstractCouchBaseSetOperator;
/**
* <p>CouchBaseOutputOperator class.</p>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java
index f2582bd..c3248bf 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import java.util.ArrayList;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java b/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java
index 3342771..d5af7a5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import java.util.HashMap;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java b/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java
index 7d02de2..7224df3 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import javax.validation.constraints.Min;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java
index 8c55404..1138e84 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
/*
* To change this template, choose Tools | Templates and open the template in the editor.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java
index 0a880fd..3282e32 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java
@@ -16,16 +16,16 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.aerospike;
+package org.apache.apex.benchmark.aerospike;
+import org.apache.apex.malhar.contrib.aerospike.AerospikeTransactionalStore;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.aerospike.AerospikeTransactionalStore;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
/**
* Application to benchmark the performance of aerospike output operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java
index f9ee689..bb3661b 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java
@@ -16,16 +16,16 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.aerospike;
+package org.apache.apex.benchmark.aerospike;
import java.util.List;
+import org.apache.apex.malhar.contrib.aerospike.AbstractAerospikeTransactionalPutOperator;
+
import com.aerospike.client.AerospikeException;
import com.aerospike.client.Bin;
import com.aerospike.client.Key;
-import com.datatorrent.contrib.aerospike.AbstractAerospikeTransactionalPutOperator;
-
/**
* <p>AerospikeOutputOperator class.</p>
*
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java
index f74311e..8e1d8b8 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java
@@ -16,8 +16,13 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.algo;
+package org.apache.apex.benchmark.algo;
+import org.apache.apex.malhar.lib.algo.UniqueCounter;
+import org.apache.apex.malhar.lib.converter.MapToKeyHashValuePairConverter;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.stream.Counter;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context;
@@ -27,13 +32,6 @@ import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
import com.datatorrent.common.partitioner.StatelessPartitioner;
-import com.datatorrent.lib.algo.UniqueCounter;
-import com.datatorrent.lib.converter.MapToKeyHashValuePairConverter;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-
-import com.datatorrent.lib.stream.Counter;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
-
/**
* Application to demonstrate PartitionableUniqueCount operator. <br>
* The input operator generate random keys, which is sent to
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java
index 46d503f..6f91cd1 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java
@@ -16,8 +16,10 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.cassandra;
+package org.apache.apex.benchmark.cassandra;
+import org.apache.apex.malhar.contrib.cassandra.CassandraTransactionalStore;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
@@ -26,10 +28,6 @@ import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.cassandra.CassandraTransactionalStore;
-
-import com.datatorrent.lib.testbench.RandomEventGenerator;
-
/**
* Application to benchmark the performance of cassandra output operator.
* The operator was tested on following configuration:
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java
index 592d8a2..90bc44b 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java
@@ -16,15 +16,15 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.cassandra;
+package org.apache.apex.benchmark.cassandra;
+
+import org.apache.apex.malhar.contrib.cassandra.AbstractCassandraTransactionableOutputOperator;
import com.datastax.driver.core.BoundStatement;
import com.datastax.driver.core.PreparedStatement;
import com.datastax.driver.core.Statement;
import com.datastax.driver.core.exceptions.DriverException;
-import com.datatorrent.contrib.cassandra.AbstractCassandraTransactionableOutputOperator;
-
/**
* <p>CassandraOutputOperator class.</p>
*
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java
index ce0821c..56dd731 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java
@@ -16,13 +16,13 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.fs;
+package org.apache.apex.benchmark.fs;
import java.util.Arrays;
import javax.validation.constraints.Min;
-import com.datatorrent.lib.io.fs.AbstractFileOutputOperator;
+import org.apache.apex.malhar.lib.io.fs.AbstractFileOutputOperator;
/**
* This output operator receives
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java
index 7a63d18..baddbc1 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java
@@ -16,8 +16,10 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.fs;
+package org.apache.apex.benchmark.fs;
+import org.apache.apex.malhar.lib.counters.BasicCounters;
+import org.apache.apex.malhar.lib.testbench.RandomWordGenerator;
import org.apache.commons.lang.mutable.MutableLong;
import org.apache.hadoop.conf.Configuration;
@@ -29,10 +31,6 @@ import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.counters.BasicCounters;
-
-import com.datatorrent.lib.testbench.RandomWordGenerator;
-
/**
* Application used to benchmark HDFS output operator
* The DAG consists of random word generator operator that is
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java
index 95fa961..2ecb2eb 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
import java.sql.SQLException;
import java.sql.Statement;
@@ -28,6 +28,9 @@ import javax.validation.constraints.Min;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.hive.AbstractFSRollingOutputOperator;
+import org.apache.apex.malhar.hive.HiveOperator;
+import org.apache.apex.malhar.hive.HiveStore;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.OperatorContext;
@@ -39,9 +42,6 @@ import com.datatorrent.api.InputOperator;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.hive.AbstractFSRollingOutputOperator;
-import com.datatorrent.contrib.hive.HiveOperator;
-import com.datatorrent.contrib.hive.HiveStore;
/**
* Application used to benchmark HIVE Insert operator
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java
index 98d9ce3..800fa5a 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
import java.sql.SQLException;
import java.sql.Statement;
@@ -27,18 +27,17 @@ import java.util.Map;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.benchmark.RandomMapOutput;
+import org.apache.apex.malhar.hive.AbstractFSRollingOutputOperator;
+import org.apache.apex.malhar.hive.HiveOperator;
+import org.apache.apex.malhar.hive.HiveStore;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
import com.datatorrent.api.DAG;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.benchmark.RandomMapOutput;
-
-import com.datatorrent.contrib.hive.AbstractFSRollingOutputOperator;
-import com.datatorrent.contrib.hive.HiveOperator;
-import com.datatorrent.contrib.hive.HiveStore;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
/**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java
index e147ad7..0bde5ed 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java
@@ -16,10 +16,11 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
+
+import org.apache.apex.malhar.contrib.kafka.AbstractKafkaInputOperator;
import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.contrib.kafka.AbstractKafkaInputOperator;
import kafka.message.Message;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java
index 6353c37..37b4d0a 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
import java.util.ArrayList;
import java.util.Collection;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java
index ead6c66..e51da05 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java
@@ -16,11 +16,13 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
import java.util.Properties;
-
+import org.apache.apex.malhar.contrib.kafka.HighlevelKafkaConsumer;
+import org.apache.apex.malhar.contrib.kafka.KafkaConsumer;
+import org.apache.apex.malhar.contrib.kafka.SimpleKafkaConsumer;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.OperatorContext;
@@ -35,9 +37,6 @@ import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.contrib.kafka.HighlevelKafkaConsumer;
-import com.datatorrent.contrib.kafka.KafkaConsumer;
-import com.datatorrent.contrib.kafka.SimpleKafkaConsumer;
/**
* The stream app to test the benckmark of kafka
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java
index 0dd4352..e8578a6 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
import org.apache.hadoop.conf.Configuration;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java
index 65601d5..64ef2c7 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
import kafka.producer.Partitioner;
import kafka.utils.VerifiableProperties;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java b/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java
index 5ee1888..9f99b8f 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java
@@ -19,4 +19,4 @@
/**
* Apex malhar performance demonstration application.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java
index b86cd01..065f546 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java
@@ -16,8 +16,12 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.script;
+package org.apache.apex.benchmark.script;
+import org.apache.apex.benchmark.RandomMapOutput;
+import org.apache.apex.malhar.contrib.ruby.RubyOperator;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -27,12 +31,6 @@ import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.benchmark.RandomMapOutput;
-import com.datatorrent.contrib.ruby.RubyOperator;
-
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
-
/**
*
* Application to benchmark the performance of ruby operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java
index 300ea64..4419a11 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java
@@ -16,8 +16,9 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
import org.apache.apex.malhar.lib.state.spillable.managed.ManagedStateSpillableStateStore;
import org.apache.hadoop.conf.Configuration;
@@ -26,7 +27,6 @@ import com.google.common.base.Preconditions;
import com.datatorrent.api.DAG;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.fileaccess.TFileImpl;
@ApplicationAnnotation(name = "SpillableBenchmarkApp")
/**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java
index 9bee9a6..0285f04 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.InputOperator;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java
index 7c45106..c827acb 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java
index 2dc6f0d..5ba8a33 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.state;
+package org.apache.apex.benchmark.state;
import java.io.IOException;
import java.io.Serializable;
@@ -27,7 +27,9 @@ import org.joda.time.Duration;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
import org.apache.apex.malhar.lib.state.managed.ManagedTimeUnifiedStateImpl;
+import org.apache.apex.malhar.lib.util.KeyValPair;
import org.apache.hadoop.conf.Configuration;
import com.google.common.collect.Lists;
@@ -43,8 +45,6 @@ import com.datatorrent.api.StatsListener;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.fileaccess.TFileImpl;
-import com.datatorrent.lib.util.KeyValPair;
@ApplicationAnnotation(name = "ManagedStateBenchmark")
/**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java
index 60a775c..eea4ac5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.state;
+package org.apache.apex.benchmark.state;
import java.nio.ByteBuffer;
import java.util.LinkedList;
@@ -28,6 +28,7 @@ import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.apache.apex.malhar.lib.state.managed.ManagedTimeUnifiedStateImpl;
+import org.apache.apex.malhar.lib.util.KeyValPair;
import com.google.common.collect.Maps;
@@ -35,7 +36,6 @@ import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.api.Operator;
import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.util.KeyValPair;
import com.datatorrent.netlet.util.Slice;
/**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java
index b0b7314..ef853e1 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java
@@ -16,8 +16,9 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
+import org.apache.apex.malhar.lib.stream.DevNullCounter;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -25,11 +26,10 @@ import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNullCounter;
/**
*
- * Functional tests for {@link com.datatorrent.lib.testbench.DevNullCounter}.
+ * Functional tests for {@link org.apache.apex.malhar.lib.testbench.DevNullCounter}.
* <p>
* <br>
* oper.process is called a billion times<br>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java
index c716206..1fca7e5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultOutputPort;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java
index 2e5bcf9..37936e6 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java
@@ -16,8 +16,10 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.stream.StreamDuplicater;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -25,8 +27,6 @@ import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.stream.StreamDuplicater;
/**
* Benchmark App for StreamDuplicater Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java
index bb1d081..de9507a 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java
@@ -16,8 +16,10 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
+import org.apache.apex.benchmark.WordCountOperator;
+import org.apache.apex.malhar.lib.stream.StreamMerger;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -25,8 +27,6 @@ import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.benchmark.WordCountOperator;
-import com.datatorrent.lib.stream.StreamMerger;
/**
* Benchmark App for StreamMerge Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java
index b1ddbee..d12267e 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java
@@ -16,11 +16,13 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.ArrayList;
import java.util.HashMap;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.EventClassifier;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -28,8 +30,6 @@ import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.EventClassifier;
/**
* Benchmark App for EventClassifier Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java
index 5fe478b..c1405fc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java
@@ -16,10 +16,13 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.HashMap;
+import org.apache.apex.benchmark.WordCountOperator;
+import org.apache.apex.benchmark.stream.IntegerOperator;
+import org.apache.apex.malhar.lib.testbench.EventClassifierNumberToHashDouble;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -27,9 +30,6 @@ import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.benchmark.WordCountOperator;
-import com.datatorrent.benchmark.stream.IntegerOperator;
-import com.datatorrent.lib.testbench.EventClassifierNumberToHashDouble;
/**
* Benchmark App for EventClassifierNumberToHashDouble Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java
index 8f28ae6..0c997dc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java
@@ -16,10 +16,12 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.HashMap;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.EventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -27,8 +29,6 @@ import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.EventGenerator;
/**
* Benchmark App for EventGenerator Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java
index e562224..d7aa2dc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java
@@ -16,19 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.ArrayList;
import java.util.HashMap;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.EventIncrementer;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.EventIncrementer;
/**
* Benchmark App for EventIncrementer Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java
index ea2943f..57664ac 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java
@@ -16,19 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.ArrayList;
import java.util.HashMap;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.FilterClassifier;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.FilterClassifier;
/**
* Benchmark App for FilterClassifier Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java
index 52c0bed..f282838 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java
@@ -16,19 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.ArrayList;
import java.util.HashMap;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.FilteredEventClassifier;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.FilteredEventClassifier;
/**
* Benchmark App for FilteredEventClassifier Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java
index 29cd079..66fab36 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.ArrayList;
import java.util.HashMap;
@@ -24,10 +24,11 @@ import java.util.HashMap;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.testbench.EventGenerator;
+
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.InputOperator;
-import com.datatorrent.lib.testbench.EventGenerator;
/**
* HashMap Input Operator used as a helper in testbench benchmarking apps.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java
index df5b11e..bdca047 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java
@@ -16,16 +16,16 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
/**
* Benchmark App for RandomEventGenerator Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java
index faafcbf..0b7f7b5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java
@@ -16,10 +16,14 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.ArrayList;
import java.util.HashMap;
+
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.SeedEventGenerator;
+import org.apache.apex.malhar.lib.util.KeyValPair;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.PortContext;
@@ -27,9 +31,6 @@ import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.SeedEventGenerator;
-import com.datatorrent.lib.util.KeyValPair;
/**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java
index d6e762e..4ad35f9 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java
@@ -16,18 +16,18 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.util.HashMap;
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.ThroughputCounter;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.ThroughputCounter;
/**
* Benchmark App for ThroughputCounter Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java
index bfdc0ec..a12354e 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
import java.nio.ByteBuffer;
import java.util.Random;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java
index 7a345fe..dc0ad48 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
import java.io.IOException;
import java.io.Serializable;
@@ -25,10 +25,13 @@ import org.joda.time.Duration;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.benchmark.window.WindowedOperatorBenchmarkApp.WindowedGenerator;
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
import org.apache.apex.malhar.lib.state.managed.UnboundedTimeBucketAssigner;
import org.apache.apex.malhar.lib.state.spillable.SpillableComplexComponentImpl;
import org.apache.apex.malhar.lib.state.spillable.SpillableStateStore;
import org.apache.apex.malhar.lib.state.spillable.managed.ManagedTimeUnifiedStateSpillableStateStore;
+import org.apache.apex.malhar.lib.stream.DevNull;
import org.apache.apex.malhar.lib.window.Accumulation;
import org.apache.apex.malhar.lib.window.TriggerOption;
import org.apache.apex.malhar.lib.window.WindowOption;
@@ -47,9 +50,6 @@ import com.datatorrent.api.Operator;
import com.datatorrent.api.Stats.OperatorStats;
import com.datatorrent.api.StatsListener;
import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.benchmark.window.WindowedOperatorBenchmarkApp.WindowedGenerator;
-import com.datatorrent.lib.fileaccess.TFileImpl;
-import com.datatorrent.lib.stream.DevNull;
/**
* @since 3.7.0
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java
index 7e38584..b8dce7f 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java
@@ -16,14 +16,16 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
import org.apache.apex.malhar.lib.state.spillable.SpillableComplexComponentImpl;
import org.apache.apex.malhar.lib.state.spillable.SpillableSetMultimapImpl;
import org.apache.apex.malhar.lib.state.spillable.managed.ManagedStateSpillableStateStore;
+import org.apache.apex.malhar.lib.util.KeyValPair;
import org.apache.apex.malhar.lib.utils.serde.GenericSerde;
import org.apache.apex.malhar.lib.window.Accumulation;
import org.apache.apex.malhar.lib.window.Tuple;
@@ -39,8 +41,6 @@ import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.fileaccess.TFileImpl;
-import com.datatorrent.lib.util.KeyValPair;
/**
* @since 3.7.0
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java
index 7e402fc..15a577e 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java
index cd8a3ec..644cd01 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import java.io.IOException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java
index 0a21a7c..c5249a2 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import java.io.IOException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java
index 6a1c968..aef17b3 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java
index e2936fe..3417b58 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java
@@ -16,17 +16,17 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.accumulo;
+package org.apache.apex.benchmark.accumulo;
import org.apache.accumulo.core.data.Mutation;
+import org.apache.apex.malhar.contrib.accumulo.AbstractAccumuloOutputOperator;
+import org.apache.apex.malhar.contrib.accumulo.AccumuloRowTupleGenerator;
+import org.apache.apex.malhar.contrib.accumulo.AccumuloTestHelper;
+import org.apache.apex.malhar.contrib.accumulo.AccumuloTuple;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.contrib.accumulo.AbstractAccumuloOutputOperator;
-import com.datatorrent.contrib.accumulo.AccumuloRowTupleGenerator;
-import com.datatorrent.contrib.accumulo.AccumuloTestHelper;
-import com.datatorrent.contrib.accumulo.AccumuloTuple;
/**
* BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java
index 8b47a9b..0f48d16 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.accumulo;
+package org.apache.apex.benchmark.accumulo;
import org.junit.Test;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java
index 14fe441..9eea1b7 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.aerospike;
+package org.apache.apex.benchmark.aerospike;
import org.junit.Test;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java
index 079d073..56c0cf2 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.algo;
+package org.apache.apex.benchmark.algo;
import org.junit.Test;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java
index ec4f308..1708eaf 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.cassandra;
+package org.apache.apex.benchmark.cassandra;
import org.junit.Test;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java
index 32a4907..94bc6a9 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.hbase;
+package org.apache.apex.benchmark.hbase;
import org.junit.Test;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java
index b61f1d3..b6ec667 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java
@@ -16,16 +16,16 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.hbase;
+package org.apache.apex.benchmark.hbase;
+import org.apache.apex.malhar.contrib.hbase.HBaseCsvMappingPutOperator;
+import org.apache.apex.malhar.contrib.hbase.HBaseRowStringGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.hbase.HBaseCsvMappingPutOperator;
-import com.datatorrent.contrib.hbase.HBaseRowStringGenerator;
/**
* BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java
index 653c6f6..aa50e45 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java
index e0097c6..6bd2edf 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java
index 6cb901a..8f2c42c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java
index 4de7193..1823a00 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java
index 9201cd5..413438a 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java
@@ -16,18 +16,18 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.contrib.memsql.MemsqlInputOperator;
+import org.apache.apex.malhar.lib.stream.DevNull;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.memsql.MemsqlInputOperator;
-import com.datatorrent.lib.stream.DevNull;
/**
* BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java
index a596903..f3127a5 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
import java.io.FileInputStream;
import java.io.IOException;
@@ -28,6 +28,9 @@ import org.junit.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.contrib.memsql.AbstractMemsqlOutputOperatorTest;
+import org.apache.apex.malhar.contrib.memsql.MemsqlPOJOOutputOperator;
+import org.apache.apex.malhar.contrib.memsql.MemsqlStore;
import org.apache.commons.io.IOUtils;
import org.apache.hadoop.conf.Configuration;
@@ -35,15 +38,12 @@ import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DAG;
import com.datatorrent.api.LocalMode;
import com.datatorrent.api.Operator.ProcessingMode;
-import com.datatorrent.contrib.memsql.AbstractMemsqlOutputOperatorTest;
-import com.datatorrent.contrib.memsql.MemsqlPOJOOutputOperator;
-import com.datatorrent.contrib.memsql.MemsqlStore;
import com.datatorrent.netlet.util.DTThrowable;
-import static com.datatorrent.contrib.memsql.AbstractMemsqlOutputOperatorTest.BATCH_SIZE;
-import static com.datatorrent.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.APP_ID;
-import static com.datatorrent.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.OPERATOR_ID;
-import static com.datatorrent.lib.helper.OperatorContextTestHelper.mockOperatorContext;
+import static org.apache.apex.malhar.contrib.memsql.AbstractMemsqlOutputOperatorTest.BATCH_SIZE;
+import static org.apache.apex.malhar.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.APP_ID;
+import static org.apache.apex.malhar.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.OPERATOR_ID;
+import static org.apache.apex.malhar.lib.helper.OperatorContextTestHelper.mockOperatorContext;
public class MemsqlInputBenchmarkTest
{
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java
index 297bc6d..0b4f126 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java
@@ -16,19 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.contrib.memsql.MemsqlPOJOOutputOperator;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.memsql.MemsqlPOJOOutputOperator;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
/**
* BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java
index bf82ab3..3d4766e 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
@@ -27,12 +27,12 @@ import org.junit.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.contrib.memsql.AbstractMemsqlOutputOperatorTest;
+import org.apache.apex.malhar.contrib.memsql.MemsqlStore;
import org.apache.commons.io.IOUtils;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.LocalMode;
-import com.datatorrent.contrib.memsql.AbstractMemsqlOutputOperatorTest;
-import com.datatorrent.contrib.memsql.MemsqlStore;
import com.datatorrent.netlet.util.DTThrowable;
public class MemsqlOutputBenchmarkTest
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java
index d270e7f..904ea6b 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.script;
+package org.apache.apex.benchmark.script;
import org.junit.Test;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java
index cd2c640..6490aa9 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
import java.io.File;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java
index b87fec1..f6ecf0c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
import java.util.Random;
@@ -27,14 +27,13 @@ import org.junit.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
import org.apache.apex.malhar.lib.state.spillable.SpillableMapImpl;
import org.apache.apex.malhar.lib.state.spillable.SpillableTestUtils;
import org.apache.apex.malhar.lib.state.spillable.managed.ManagedStateSpillableStateStore;
import org.apache.apex.malhar.lib.utils.serde.Serde;
import org.apache.apex.malhar.lib.utils.serde.StringSerde;
-import com.datatorrent.lib.fileaccess.TFileImpl;
-
public class SpillableDSBenchmarkTest
{
private static final Logger logger = LoggerFactory.getLogger(SpillableDSBenchmarkTest.class);
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java
index dc8f4b4..eb3cc7a 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java
@@ -16,20 +16,20 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.state;
+package org.apache.apex.benchmark.state;
import java.io.File;
import org.junit.Before;
import org.junit.Test;
+import org.apache.apex.benchmark.state.StoreOperator.ExecMode;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileUtil;
import com.datatorrent.api.DAG;
import com.datatorrent.api.LocalMode;
import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.benchmark.state.StoreOperator.ExecMode;
/**
* This is not a really unit test, but in fact a benchmark runner.
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java
index 99d8a1f..88d66b4 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java
index 929d8bc..9fb4638 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java
index 5a427a5..0a71ea4 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java
index 1a85a7b..fb4401e 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java
index 9419022..c921a4c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java
index 977d6b7..658e372 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java
index 92ca0fd..b278f6d 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java
index 157accc..64fca9e 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.util.serde;
+package org.apache.apex.benchmark.util.serde;
import java.util.Random;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java
index 2bc9335..fe9ba0c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
import java.io.File;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java
index 4a16396..b6f3549 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
import java.io.File;
diff --git a/contrib/pom.xml b/contrib/pom.xml
index a69ad29..13c77b9 100755
--- a/contrib/pom.xml
+++ b/contrib/pom.xml
@@ -25,7 +25,7 @@
<parent>
<groupId>org.apache.apex</groupId>
<artifactId>malhar</artifactId>
- <version>3.9.0-SNAPSHOT</version>
+ <version>4.0.0-SNAPSHOT</version>
</parent>
<artifactId>malhar-contrib</artifactId>
@@ -36,7 +36,6 @@
<!-- skip tests by default as they depend on manual setup -->
<skip.contrib.module.tests>false</skip.contrib.module.tests>
<skipTests>true</skipTests>
- <checkstyle.console>false</checkstyle.console>
</properties>
<repositories>
<repository>
@@ -217,14 +216,6 @@
<artifactId>japicmp-maven-plugin</artifactId>
</plugin>
- <plugin>
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-checkstyle-plugin</artifactId>
- <configuration>
- <maxAllowedViolations>2709</maxAllowedViolations>
- <logViolationsToConsole>${checkstyle.console}</logViolationsToConsole>
- </configuration>
- </plugin>
</plugins>
</build>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java
index 50a1056..8250cd0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
import java.util.Map.Entry;
@@ -24,8 +24,8 @@ import org.apache.accumulo.core.client.Connector;
import org.apache.accumulo.core.client.Scanner;
import org.apache.accumulo.core.data.Key;
import org.apache.accumulo.core.data.Value;
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
/**
* Base input adapter, which reads data from persistence database and writes into output port(s). Subclasses should provide the
* implementation of getting the tuples and scanner methods.
@@ -38,18 +38,21 @@ import com.datatorrent.lib.db.AbstractStoreInputOperator;
* @tags key value, accumulo
* @since 1.0.4
*/
-public abstract class AbstractAccumuloInputOperator<T> extends AbstractStoreInputOperator<T, AccumuloStore> {
+public abstract class AbstractAccumuloInputOperator<T> extends AbstractStoreInputOperator<T, AccumuloStore>
+{
public abstract T getTuple(Entry<Key, Value> entry);
public abstract Scanner getScanner(Connector conn);
- public AbstractAccumuloInputOperator() {
+ public AbstractAccumuloInputOperator()
+ {
store = new AccumuloStore();
}
@Override
- public void emitTuples() {
+ public void emitTuples()
+ {
Connector conn = store.getConnector();
Scanner scan = getScanner(conn);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java
index a4ae1ac..153f88d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java
@@ -16,19 +16,21 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
import java.util.List;
-import org.apache.accumulo.core.client.MutationsRejectedException;
-import org.apache.accumulo.core.data.Mutation;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.accumulo.core.client.MutationsRejectedException;
+import org.apache.accumulo.core.data.Mutation;
+import org.apache.apex.malhar.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
+
+import com.google.common.collect.Lists;
+
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
-import com.google.common.collect.Lists;
/**
* Base output operator that stores tuples in Accumulo rows. Subclasses should provide implementation of operationMutation method. <br>
@@ -56,7 +58,8 @@ import com.google.common.collect.Lists;
* The tuple type
* @since 1.0.4
*/
-public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregateTransactionableStoreOutputOperator<T, AccumuloWindowStore> {
+public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregateTransactionableStoreOutputOperator<T, AccumuloWindowStore>
+{
private static final transient Logger logger = LoggerFactory.getLogger(AbstractAccumuloOutputOperator.class);
private final List<T> tuples;
private transient ProcessingMode mode;
@@ -75,6 +78,7 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
tuples = Lists.newArrayList();
store = new AccumuloWindowStore();
}
+
@Override
public void processTuple(T tuple)
{
@@ -82,7 +86,8 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
}
@Override
- public void storeAggregate() {
+ public void storeAggregate()
+ {
try {
for (T tuple : tuples) {
Mutation mutation = operationMutation(tuple);
@@ -96,6 +101,7 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
}
tuples.clear();
}
+
/**
*
* @param t
@@ -106,11 +112,11 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
@Override
public void setup(OperatorContext context)
{
- mode=context.getValue(context.PROCESSING_MODE);
- if(mode==ProcessingMode.EXACTLY_ONCE){
+ mode = context.getValue(context.PROCESSING_MODE);
+ if (mode == ProcessingMode.EXACTLY_ONCE) {
throw new RuntimeException("This operator only supports atmost once and atleast once processing modes");
}
- if(mode==ProcessingMode.AT_MOST_ONCE){
+ if (mode == ProcessingMode.AT_MOST_ONCE) {
tuples.clear();
}
super.setup(context);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java
index 52d273a..2eb4db0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java
@@ -16,10 +16,13 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
import java.io.IOException;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
import org.apache.accumulo.core.client.AccumuloException;
import org.apache.accumulo.core.client.AccumuloSecurityException;
import org.apache.accumulo.core.client.BatchWriter;
@@ -30,11 +33,9 @@ import org.apache.accumulo.core.client.MutationsRejectedException;
import org.apache.accumulo.core.client.TableNotFoundException;
import org.apache.accumulo.core.client.ZooKeeperInstance;
import org.apache.accumulo.core.client.security.tokens.PasswordToken;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.Connectable;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.Connectable;
/**
* A {@link Connectable} for accumulo that implements Connectable interface.
@@ -45,7 +46,8 @@ import com.datatorrent.lib.db.Connectable;
* @param <T>
* @since 1.0.4
*/
-public class AccumuloStore implements Connectable {
+public class AccumuloStore implements Connectable
+{
private static final transient Logger logger = LoggerFactory.getLogger(AccumuloStore.class);
private String zookeeperHost;
private String instanceName;
@@ -58,12 +60,13 @@ public class AccumuloStore implements Connectable {
private long memoryLimit;
private int numThreads;
- private static final long DEFAULT_MEMORY=2147483648l;
- private static final int DEFAULT_THREADS=1;
+ private static final long DEFAULT_MEMORY = 2147483648L;
+ private static final int DEFAULT_THREADS = 1;
- public AccumuloStore(){
- memoryLimit=DEFAULT_MEMORY;
- numThreads=DEFAULT_THREADS;
+ public AccumuloStore()
+ {
+ memoryLimit = DEFAULT_MEMORY;
+ numThreads = DEFAULT_THREADS;
}
/**
@@ -71,7 +74,8 @@ public class AccumuloStore implements Connectable {
*
* @return Connector
*/
- public Connector getConnector() {
+ public Connector getConnector()
+ {
return connector;
}
@@ -80,7 +84,8 @@ public class AccumuloStore implements Connectable {
*
* @return TableName
*/
- public String getTableName() {
+ public String getTableName()
+ {
return tableName;
}
@@ -89,7 +94,8 @@ public class AccumuloStore implements Connectable {
*
* @param tableName
*/
- public void setTableName(String tableName) {
+ public void setTableName(String tableName)
+ {
this.tableName = tableName;
}
@@ -98,7 +104,8 @@ public class AccumuloStore implements Connectable {
*
* @return ZookeeperHost
*/
- public String getZookeeperHost() {
+ public String getZookeeperHost()
+ {
return zookeeperHost;
}
@@ -107,7 +114,8 @@ public class AccumuloStore implements Connectable {
*
* @param zookeeperHost
*/
- public void setZookeeperHost(String zookeeperHost) {
+ public void setZookeeperHost(String zookeeperHost)
+ {
this.zookeeperHost = zookeeperHost;
}
@@ -116,7 +124,8 @@ public class AccumuloStore implements Connectable {
*
* @return instanceName
*/
- public String getInstanceName() {
+ public String getInstanceName()
+ {
return instanceName;
}
@@ -125,7 +134,8 @@ public class AccumuloStore implements Connectable {
*
* @param instanceName
*/
- public void setInstanceName(String instanceName) {
+ public void setInstanceName(String instanceName)
+ {
this.instanceName = instanceName;
}
@@ -134,7 +144,8 @@ public class AccumuloStore implements Connectable {
*
* @param userName
*/
- public void setUserName(String userName) {
+ public void setUserName(String userName)
+ {
this.userName = userName;
}
@@ -143,23 +154,28 @@ public class AccumuloStore implements Connectable {
*
* @param password
*/
- public void setPassword(String password) {
+ public void setPassword(String password)
+ {
this.password = password;
}
+
/**
* setter for memory limit
*
* @param memoryLimit
*/
- public void setMemoryLimit(long memoryLimit) {
+ public void setMemoryLimit(long memoryLimit)
+ {
this.memoryLimit = memoryLimit;
}
+
/**
* setter for number of writer threads
*
* @param numThreads
*/
- public void setNumThreads(int numThreads) {
+ public void setNumThreads(int numThreads)
+ {
this.numThreads = numThreads;
}
@@ -168,12 +184,14 @@ public class AccumuloStore implements Connectable {
*
* @return BatchWriter
*/
- public BatchWriter getBatchwriter() {
+ public BatchWriter getBatchwriter()
+ {
return batchwriter;
}
@Override
- public void connect() throws IOException {
+ public void connect() throws IOException
+ {
Instance instance = null;
instance = new ZooKeeperInstance(instanceName, zookeeperHost);
try {
@@ -198,7 +216,8 @@ public class AccumuloStore implements Connectable {
}
@Override
- public void disconnect() throws IOException {
+ public void disconnect() throws IOException
+ {
try {
batchwriter.close();
} catch (MutationsRejectedException e) {
@@ -208,7 +227,8 @@ public class AccumuloStore implements Connectable {
}
@Override
- public boolean isConnected() {
+ public boolean isConnected()
+ {
// Not applicable for accumulo
return false;
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java
index 9bec031..412ae14 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
@@ -25,6 +25,9 @@ import java.io.DataOutputStream;
import java.io.IOException;
import java.util.Map.Entry;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
import org.apache.accumulo.core.client.MutationsRejectedException;
import org.apache.accumulo.core.client.Scanner;
import org.apache.accumulo.core.client.TableNotFoundException;
@@ -33,12 +36,10 @@ import org.apache.accumulo.core.data.Mutation;
import org.apache.accumulo.core.data.Range;
import org.apache.accumulo.core.data.Value;
import org.apache.accumulo.core.security.Authorizations;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
import org.apache.hadoop.io.Text;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.TransactionableStore;
/**
* Provides transactional support by implementing TransactionableStore abstract methods.
@@ -53,7 +54,8 @@ import com.datatorrent.lib.db.TransactionableStore;
* @tags accumulo, key value
* @since 1.0.4
*/
-public class AccumuloWindowStore extends AccumuloStore implements TransactionableStore {
+public class AccumuloWindowStore extends AccumuloStore implements TransactionableStore
+{
private static final transient Logger logger = LoggerFactory.getLogger(AccumuloWindowStore.class);
private static final String DEFAULT_ROW_NAME = "AccumuloOperator_row";
private static final String DEFAULT_COLUMN_FAMILY_NAME = "AccumuloOutputOperator_cf";
@@ -67,7 +69,8 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
private transient String lastWindowColumnName;
private transient byte[] lastWindowColumnBytes;
- public AccumuloWindowStore() {
+ public AccumuloWindowStore()
+ {
rowName = DEFAULT_ROW_NAME;
columnFamilyName = DEFAULT_COLUMN_FAMILY_NAME;
lastWindowColumnName = DEFAULT_LAST_WINDOW_PREFIX_COLUMN_NAME;
@@ -78,54 +81,64 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
* the values are stored as byte arrays.This method converts string to byte
* arrays. uses util class in hbase library to do so.
*/
- private void constructKeys() {
+ private void constructKeys()
+ {
rowBytes = rowName.getBytes();
columnFamilyBytes = columnFamilyName.getBytes();
}
- public String getRowName() {
+ public String getRowName()
+ {
return rowName;
}
- public void setRowName(String rowName) {
+ public void setRowName(String rowName)
+ {
this.rowName = rowName;
constructKeys();
}
- public String getColumnFamilyName() {
+ public String getColumnFamilyName()
+ {
return columnFamilyName;
}
- public void setColumnFamilyName(String columnFamilyName) {
+ public void setColumnFamilyName(String columnFamilyName)
+ {
this.columnFamilyName = columnFamilyName;
constructKeys();
}
@Override
- public void beginTransaction() {
+ public void beginTransaction()
+ {
// accumulo does not support transactions
}
@Override
- public void commitTransaction() {
+ public void commitTransaction()
+ {
// accumulo does not support transactions
}
@Override
- public void rollbackTransaction() {
+ public void rollbackTransaction()
+ {
// accumulo does not support transactions
}
@Override
- public boolean isInTransaction() {
+ public boolean isInTransaction()
+ {
// accumulo does not support transactions
return false;
}
@Override
- public long getCommittedWindowId(String appId, int operatorId) {
+ public long getCommittedWindowId(String appId, int operatorId)
+ {
byte[] value = null;
Authorizations auths = new Authorizations();
Scanner scan = null;
@@ -150,7 +163,8 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
}
@Override
- public void storeCommittedWindowId(String appId, int operatorId,long windowId) {
+ public void storeCommittedWindowId(String appId, int operatorId,long windowId)
+ {
byte[] WindowIdBytes = toBytes(windowId);
String columnKey = appId + "_" + operatorId + "_" + lastWindowColumnName;
lastWindowColumnBytes = columnKey.getBytes();
@@ -166,19 +180,19 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
}
@Override
- public void removeCommittedWindowId(String appId, int operatorId) {
+ public void removeCommittedWindowId(String appId, int operatorId)
+ {
// accumulo does not support transactions
-
}
public static byte[] toBytes(long l)
{
- ByteArrayOutputStream baos=new ByteArrayOutputStream(Long.SIZE/8);
- DataOutputStream dos=new DataOutputStream(baos);
- byte[] result=null;
+ ByteArrayOutputStream baos = new ByteArrayOutputStream(Long.SIZE / 8);
+ DataOutputStream dos = new DataOutputStream(baos);
+ byte[] result = null;
try {
dos.writeLong(l);
- result=baos.toByteArray();
+ result = baos.toByteArray();
dos.close();
} catch (IOException e) {
logger.error("error converting to byte array");
@@ -187,11 +201,11 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
return result;
}
-
- public static long toLong(byte[] b){
- ByteArrayInputStream baos=new ByteArrayInputStream(b);
- DataInputStream dos=new DataInputStream(baos);
- long result=0;
+ public static long toLong(byte[] b)
+ {
+ ByteArrayInputStream baos = new ByteArrayInputStream(b);
+ DataInputStream dos = new DataInputStream(baos);
+ long result = 0;
try {
result = dos.readLong();
dos.close();
@@ -202,5 +216,4 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
return result;
}
-
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java
index cbac406..508c16e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java
@@ -17,4 +17,4 @@
* under the License.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java
index 9f7469f..3cc0d69 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java
@@ -16,31 +16,34 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
import com.aerospike.client.Record;
import com.aerospike.client.query.RecordSet;
import com.aerospike.client.query.Statement;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
- * Base input adapter, which reads data from persistence database through its API and writes into output port(s).&nsbsp;
- * Subclasses should provide the implementation of getting the tuples and querying to retrieve data.
+ * Base input adapter, which reads data from persistence database through its
+ * API and writes into output port(s). Subclasses should provide the
+ * implementation of getting the tuples and querying to retrieve data.
* <p>
- * This is an abstract class. Sub-classes need to implement {@link #queryToRetrieveData()} and {@link #getTuple(Record)}.
- * </p>
+ * This is an abstract class. Sub-classes need to implement
+ * {@link #queryToRetrieveData()} and {@link #getTuple(Record)}.
+ *
* @displayName Abstract Aerospike Get
* @category Input
* @tags get
* @since 1.0.4
*/
-public abstract class AbstractAerospikeGetOperator<T> extends AbstractStoreInputOperator<T, AerospikeStore> {
-
+public abstract class AbstractAerospikeGetOperator<T> extends AbstractStoreInputOperator<T, AerospikeStore>
+{
private static final Logger logger = LoggerFactory.getLogger(AbstractAerospikeGetOperator.class);
/**
@@ -69,20 +72,19 @@ public abstract class AbstractAerospikeGetOperator<T> extends AbstractStoreInput
* It then converts each row into tuple and emit that into output port.
*/
@Override
- public void emitTuples() {
-
+ public void emitTuples()
+ {
Statement query = queryToRetrieveData();
logger.debug(String.format("select statement: %s", query.toString()));
RecordSet rs;
try {
rs = store.getClient().query(null, query);
- while(rs.next()){
+ while (rs.next()) {
Record rec = rs.getRecord();
T tuple = getTuple(rec);
outputPort.emit(tuple);
}
- }
- catch (Exception ex) {
+ } catch (Exception ex) {
store.disconnect();
DTThrowable.rethrow(ex);
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java
index a629078..ca14a37 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java
@@ -16,32 +16,36 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
import java.util.List;
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
+
import com.aerospike.client.AerospikeException;
import com.aerospike.client.Bin;
import com.aerospike.client.Key;
import com.google.common.collect.Lists;
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
-
/**
- * <p>
- * Generic base output adaptor which writes tuples as they come without providing any transactional support. Subclasses should provide implementation for getting updated bins.
- * </p>
+ * Generic base output adaptor which writes tuples as they come without
+ * providing any transactional support. Subclasses should provide implementation
+ * for getting updated bins.
+ *
* @displayName Abstract Aerospike Non Transactional Put
* @category Output
* @tags put, non transactional
* @param <T> type of tuple
* @since 1.0.4
*/
-public abstract class AbstractAerospikeNonTransactionalPutOperator<T> extends AbstractStoreOutputOperator<T,AerospikeStore> {
+public abstract class AbstractAerospikeNonTransactionalPutOperator<T>
+ extends AbstractStoreOutputOperator<T,AerospikeStore>
+{
+ private final transient List<Bin> bins;
- private transient final List<Bin> bins;
- public AbstractAerospikeNonTransactionalPutOperator() {
+ public AbstractAerospikeNonTransactionalPutOperator()
+ {
super();
bins = Lists.newArrayList();
}
@@ -58,8 +62,8 @@ public abstract class AbstractAerospikeNonTransactionalPutOperator<T> extends Ab
protected abstract Key getUpdatedBins(T tuple, List<Bin> bins) throws AerospikeException;
@Override
- public void processTuple(T tuple) {
-
+ public void processTuple(T tuple)
+ {
Key key;
Bin[] binsArray;
try {
@@ -68,11 +72,9 @@ public abstract class AbstractAerospikeNonTransactionalPutOperator<T> extends Ab
binsArray = bins.toArray(binsArray);
store.getClient().put(null, key, binsArray);
bins.clear();
- }
- catch (AerospikeException e) {
+ } catch (AerospikeException e) {
throw new RuntimeException(e);
}
-
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java
index 6db8896..9b8ba59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java
@@ -16,39 +16,42 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
import java.util.Collection;
import java.util.List;
+import org.apache.apex.malhar.lib.db.AbstractBatchTransactionableStoreOutputOperator;
+
import com.aerospike.client.AerospikeException;
import com.aerospike.client.Bin;
import com.aerospike.client.Key;
import com.google.common.collect.Lists;
-import com.datatorrent.lib.db.AbstractBatchTransactionableStoreOutputOperator;
-
/**
+ * Generic base adaptor which creates a transaction at the start of window.
+ * Subclasses should provide implementation for getting updated bins.
* <p>
- * Generic base adaptor which creates a transaction at the start of window. Subclasses should provide implementation for getting updated bins. <br/>
- * </p>
- * <p>
- * Executes all the put updates and closes the transaction at the end of the window.
- * The tuples in a window are stored in check-pointed collection which is cleared in the endWindow().
- * This is needed for the recovery. The operator writes a tuple at least once in the database, which is why
- * only when all the updates are executed, the transaction is committed in the end window call.
+ * Executes all the put updates and closes the transaction at the end of the
+ * window. The tuples in a window are stored in check-pointed collection which
+ * is cleared in the endWindow(). This is needed for the recovery. The operator
+ * writes a tuple at least once in the database, which is why only when all the
+ * updates are executed, the transaction is committed in the end window call.
* </p>
+ *
* @displayName Abstract Aerospike Transactional Put
* @category Output
* @tags put, transactional
* @param <T>type of tuple
* @since 1.0.4
*/
-public abstract class AbstractAerospikeTransactionalPutOperator<T> extends AbstractBatchTransactionableStoreOutputOperator<T, AerospikeTransactionalStore> {
-
- private transient final List<Bin> bins;
+public abstract class AbstractAerospikeTransactionalPutOperator<T>
+ extends AbstractBatchTransactionableStoreOutputOperator<T, AerospikeTransactionalStore>
+{
+ private final transient List<Bin> bins;
- public AbstractAerospikeTransactionalPutOperator() {
+ public AbstractAerospikeTransactionalPutOperator()
+ {
super();
bins = Lists.newArrayList();
}
@@ -70,7 +73,7 @@ public abstract class AbstractAerospikeTransactionalPutOperator<T> extends Abstr
Key key;
Bin[] binsArray;
try {
- for(T tuple: tuples) {
+ for (T tuple: tuples) {
key = getUpdatedBins(tuple,bins);
binsArray = new Bin[bins.size()];
binsArray = bins.toArray(binsArray);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java
index f69e433..49580d0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java
@@ -16,22 +16,22 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
import java.util.ArrayList;
import java.util.List;
import javax.validation.constraints.NotNull;
-import com.aerospike.client.Bin;
-import com.aerospike.client.Key;
-
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
+
+import com.aerospike.client.Bin;
+import com.aerospike.client.Key;
/**
* <p>
@@ -47,8 +47,8 @@ import com.datatorrent.lib.util.PojoUtils.Getter;
@Evolving
public class AerospikePOJONonTransactionalPutOperator extends AbstractAerospikeNonTransactionalPutOperator<Object>
{
- private static transient final Logger LOG
- = LoggerFactory.getLogger(AerospikePOJONonTransactionalPutOperator.class);
+ private static final Logger LOG =
+ LoggerFactory.getLogger(AerospikePOJONonTransactionalPutOperator.class);
// Two element list; first retrieves the record key and second the list of bins in this tuple
@NotNull
@@ -91,7 +91,7 @@ public class AerospikePOJONonTransactionalPutOperator extends AbstractAerospikeN
}
Key key = keyGetter.get(tuple);
List<Bin> binList = binsGetter.get(tuple);
- if ( ! (null == binList || binList.isEmpty()) ) {
+ if (!(null == binList || binList.isEmpty())) {
list.addAll(binList);
}
return key;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java
index f59b893..7add642 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java
@@ -16,27 +16,24 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
import java.util.ArrayList;
import java.util.List;
import javax.validation.constraints.NotNull;
-import com.aerospike.client.Bin;
-import com.aerospike.client.Key;
-
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
import org.apache.hadoop.classification.InterfaceStability.Evolving;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
+import com.aerospike.client.Bin;
+import com.aerospike.client.Key;
/**
* <p>
* A generic implementation of
- * {@link com.datatorrent.contrib.aerospike.AbstractAerospikeTransactionalPutOperator} which can
+ * {@link org.apache.apex.malhar.contrib.aerospike.AbstractAerospikeTransactionalPutOperator} which can
* take a POJO.
* </p>
* @displayName Aerospike Transactional Put Operator
@@ -45,12 +42,8 @@ import com.datatorrent.lib.util.PojoUtils.Getter;
* @since 2.1.0
*/
@Evolving
-public class AerospikePOJOTransactionalPutOperator
- extends AbstractAerospikeTransactionalPutOperator<Object>
+public class AerospikePOJOTransactionalPutOperator extends AbstractAerospikeTransactionalPutOperator<Object>
{
- private static transient final Logger LOG
- = LoggerFactory.getLogger(AerospikePOJOTransactionalPutOperator.class);
-
// Two element list; first retrieves the record key and second the list of bins in this tuple
@NotNull
private ArrayList<String> expressions;
@@ -92,11 +85,10 @@ public class AerospikePOJOTransactionalPutOperator
}
Key key = keyGetter.get(tuple);
List<Bin> binList = binsGetter.get(tuple);
- if ( ! (null == binList || binList.isEmpty()) ) {
+ if ( !(null == binList || binList.isEmpty()) ) {
list.addAll(binList);
}
return key;
}
-
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java
index 10809e1..2ffda24 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java
@@ -16,15 +16,18 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
import javax.validation.constraints.NotNull;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.Connectable;
+
import com.aerospike.client.AerospikeClient;
import com.aerospike.client.AerospikeException;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.Connectable;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* A {@link Connectable} that uses aerospike to connect to stores and implements Connectable interface.
@@ -34,8 +37,8 @@ import org.slf4j.LoggerFactory;
* @tags store
* @since 1.0.4
*/
-public class AerospikeStore implements Connectable {
-
+public class AerospikeStore implements Connectable
+{
protected static final Logger logger = LoggerFactory.getLogger(AerospikeStore.class);
private String userName;
private String password;
@@ -49,7 +52,8 @@ public class AerospikeStore implements Connectable {
*
* @param userName user name.
*/
- public void setUserName(String userName) {
+ public void setUserName(String userName)
+ {
this.userName = userName;
}
@@ -58,7 +62,8 @@ public class AerospikeStore implements Connectable {
*
* @param password password
*/
- public void setPassword(String password) {
+ public void setPassword(String password)
+ {
this.password = password;
}
@@ -68,7 +73,8 @@ public class AerospikeStore implements Connectable {
* @return The node
*/
@NotNull
- public String getNode() {
+ public String getNode()
+ {
return node;
}
@@ -77,7 +83,8 @@ public class AerospikeStore implements Connectable {
*
* @param node node
*/
- public void setNode(@NotNull String node) {
+ public void setNode(@NotNull String node)
+ {
this.node = node;
}
@@ -86,7 +93,8 @@ public class AerospikeStore implements Connectable {
*
* @return The client
*/
- public AerospikeClient getClient() {
+ public AerospikeClient getClient()
+ {
return client;
}
@@ -95,7 +103,8 @@ public class AerospikeStore implements Connectable {
*
* @param port port
*/
- public void setPort(int port) {
+ public void setPort(int port)
+ {
this.port = port;
}
@@ -103,15 +112,14 @@ public class AerospikeStore implements Connectable {
* Create connection with database.
*/
@Override
- public void connect() {
+ public void connect()
+ {
try {
client = new AerospikeClient(node, port);
logger.debug("Aerospike connection Success");
- }
- catch (AerospikeException ex) {
+ } catch (AerospikeException ex) {
throw new RuntimeException("closing database resource", ex);
- }
- catch (Throwable t) {
+ } catch (Throwable t) {
DTThrowable.rethrow(t);
}
}
@@ -120,12 +128,14 @@ public class AerospikeStore implements Connectable {
* Close connection.
*/
@Override
- public void disconnect() {
+ public void disconnect()
+ {
client.close();
}
@Override
- public boolean isConnected() {
+ public boolean isConnected()
+ {
return !client.isConnected();
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java
index 8128a17..8f898e6 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java
@@ -16,10 +16,12 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
import javax.annotation.Nonnull;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
+
import com.aerospike.client.AerospikeException;
import com.aerospike.client.Bin;
import com.aerospike.client.Key;
@@ -28,7 +30,6 @@ import com.aerospike.client.query.IndexType;
import com.aerospike.client.query.RecordSet;
import com.aerospike.client.query.Statement;
import com.aerospike.client.task.IndexTask;
-import com.datatorrent.lib.db.TransactionableStore;
/**
* <p>Provides transaction support to the operators by implementing TransactionableStore abstract methods. </p>
@@ -38,8 +39,8 @@ import com.datatorrent.lib.db.TransactionableStore;
* @tags store, transactional
* @since 1.0.4
*/
-public class AerospikeTransactionalStore extends AerospikeStore implements TransactionableStore {
-
+public class AerospikeTransactionalStore extends AerospikeStore implements TransactionableStore
+{
public static String DEFAULT_APP_ID_COL = "dt_app_id";
public static String DEFAULT_OPERATOR_ID_COL = "dt_operator_id";
public static String DEFAULT_WINDOW_COL = "dt_window";
@@ -59,8 +60,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
private transient boolean inTransaction;
private transient Statement lastWindowFetchCommand;
- public AerospikeTransactionalStore() {
-
+ public AerospikeTransactionalStore()
+ {
super();
metaSet = DEFAULT_META_SET;
metaTableAppIdColumn = DEFAULT_APP_ID_COL;
@@ -75,8 +76,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
*
* @param metaSet meta set name.
*/
- public void setMetaSet(@Nonnull String metaSet) {
-
+ public void setMetaSet(@Nonnull String metaSet)
+ {
this.metaSet = metaSet;
}
@@ -86,8 +87,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
*
* @param appIdColumn application id column name.
*/
- public void setMetaTableAppIdColumn(@Nonnull String appIdColumn) {
-
+ public void setMetaTableAppIdColumn(@Nonnull String appIdColumn)
+ {
this.metaTableAppIdColumn = appIdColumn;
}
@@ -97,8 +98,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
*
* @param operatorIdColumn operator id column name.
*/
- public void setMetaTableOperatorIdColumn(@Nonnull String operatorIdColumn) {
-
+ public void setMetaTableOperatorIdColumn(@Nonnull String operatorIdColumn)
+ {
this.metaTableOperatorIdColumn = operatorIdColumn;
}
@@ -108,8 +109,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
*
* @param windowColumn window column name.
*/
- public void setMetaTableWindowColumn(@Nonnull String windowColumn) {
-
+ public void setMetaTableWindowColumn(@Nonnull String windowColumn)
+ {
this.metaTableWindowColumn = windowColumn;
}
@@ -118,14 +119,14 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
*
* @param namespace namespace.
*/
- public void setNamespace(@Nonnull String namespace) {
-
+ public void setNamespace(@Nonnull String namespace)
+ {
this.namespace = namespace;
}
@Override
- public void connect() {
-
+ public void connect()
+ {
super.connect();
createIndexes();
try {
@@ -133,44 +134,43 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
lastWindowFetchCommand.setNamespace(namespace);
lastWindowFetchCommand.setSetName(metaSet);
lastWindowFetchCommand.setBinNames(metaTableWindowColumn);
- }
- catch (Exception e) {
+ } catch (Exception e) {
throw new RuntimeException(e);
}
}
@Override
- public void disconnect() {
-
+ public void disconnect()
+ {
super.disconnect();
}
@Override
- public void beginTransaction() {
-
+ public void beginTransaction()
+ {
inTransaction = true;
}
@Override
- public void commitTransaction() {
-
+ public void commitTransaction()
+ {
inTransaction = false;
}
@Override
- public void rollbackTransaction() {
-
+ public void rollbackTransaction()
+ {
inTransaction = false;
}
@Override
- public boolean isInTransaction() {
-
+ public boolean isInTransaction()
+ {
return inTransaction;
}
- private void createIndexes() {
-
+ private void createIndexes()
+ {
IndexTask task;
try {
task = client.createIndex(null, namespace, metaSet,
@@ -186,26 +186,25 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
}
@Override
- public long getCommittedWindowId(String appId, int operatorId) {
-
+ public long getCommittedWindowId(String appId, int operatorId)
+ {
try {
lastWindowFetchCommand.setFilters(Filter.equal(metaTableOperatorIdColumn, operatorId));
lastWindowFetchCommand.setFilters(Filter.equal(metaTableAppIdColumn, appId));
long lastWindow = -1;
RecordSet recordSet = client.query(null, lastWindowFetchCommand);
- while(recordSet.next()) {
+ while (recordSet.next()) {
lastWindow = Long.parseLong(recordSet.getRecord().getValue(metaTableWindowColumn).toString());
}
return lastWindow;
- }
- catch (AerospikeException ex) {
+ } catch (AerospikeException ex) {
throw new RuntimeException(ex);
}
}
@Override
- public void storeCommittedWindowId(String appId, int operatorId, long windowId) {
-
+ public void storeCommittedWindowId(String appId, int operatorId, long windowId)
+ {
try {
String keyString = appId + String.valueOf(operatorId);
Key key = new Key(namespace,metaSet,keyString.hashCode());
@@ -213,21 +212,19 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
Bin bin2 = new Bin(metaTableOperatorIdColumn,operatorId);
Bin bin3 = new Bin(metaTableWindowColumn,windowId);
client.put(null, key, bin1,bin2,bin3);
- }
- catch (AerospikeException e) {
+ } catch (AerospikeException e) {
throw new RuntimeException(e);
}
}
@Override
- public void removeCommittedWindowId(String appId, int operatorId) {
-
+ public void removeCommittedWindowId(String appId, int operatorId)
+ {
try {
String keyString = appId + String.valueOf(operatorId);
Key key = new Key(namespace,metaSet,keyString.hashCode());
client.delete(null, key);
- }
- catch (AerospikeException e) {
+ } catch (AerospikeException e) {
throw new RuntimeException(e);
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java
index 97e878c..8a37041 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java
@@ -17,4 +17,4 @@
* under the License.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java
index ecd57aa..7f44cc9 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java
@@ -16,18 +16,30 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.apachelog;
-
-import com.datatorrent.api.*;
-import com.datatorrent.api.Context.OperatorContext;
-import java.io.*;
+package org.apache.apex.malhar.contrib.apachelog;
+
+import java.io.BufferedReader;
+import java.io.File;
+import java.io.FileInputStream;
+import java.io.IOException;
+import java.io.InputStream;
+import java.io.InputStreamReader;
import java.text.SimpleDateFormat;
-import java.util.*;
+import java.util.ArrayList;
+import java.util.Date;
+import java.util.List;
+import java.util.Random;
+import java.util.StringTokenizer;
import java.util.concurrent.ArrayBlockingQueue;
import javax.validation.constraints.NotNull;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import com.datatorrent.api.Context.OperatorContext;
+import com.datatorrent.api.DefaultOutputPort;
+import com.datatorrent.api.InputOperator;
+import com.datatorrent.api.Operator;
+
/**
* An implementation of input operator and activation listener that simulates the apache logs.
* <p>
@@ -38,7 +50,7 @@ import org.slf4j.LoggerFactory;
*/
public class ApacheLogInputGenerator implements InputOperator, Operator.ActivationListener<OperatorContext>
{
- private final static String delimiter = ";";
+ private static final String delimiter = ";";
private transient Random random;
private transient int ipAddressCount;
@@ -99,13 +111,12 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
private List<String> readLines(String file) throws IOException
{
- List<String> lines = new ArrayList<String>();
+ List<String> lines = new ArrayList<>();
InputStream in;
File f = new File(file);
if (f.exists()) {
in = new FileInputStream(f);
- }
- else {
+ } else {
in = getClass().getResourceAsStream(file);
}
BufferedReader br = new BufferedReader(new InputStreamReader(in));
@@ -114,8 +125,7 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
while ((line = br.readLine()) != null) {
lines.add(line);
}
- }
- finally {
+ } finally {
br.close();
}
return lines;
@@ -150,8 +160,7 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
bytes.add(Integer.parseInt(token.nextToken().trim()));
status.add(Integer.parseInt(token.nextToken().trim()));
}
- }
- catch (IOException e) {
+ } catch (IOException e) {
throw new RuntimeException(e);
}
@@ -214,8 +223,7 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
if (maxDelay > 0) {
try {
Thread.sleep(random.nextInt(maxDelay));
- }
- catch (InterruptedException e) {
+ } catch (InterruptedException e) {
return;
}
}
@@ -232,8 +240,8 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
try {
thread.interrupt();
thread.join();
- }
- catch (InterruptedException ex) {
+ } catch (InterruptedException ex) {
+ // ignore
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java
index 23b4a62..b8d29d4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java
@@ -16,15 +16,16 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
+
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.io.fs.TailFsInputOperator;
+import org.apache.apex.malhar.lib.logs.ApacheLogParseMapOutputOperator;
+import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.io.fs.TailFsInputOperator;
-import com.datatorrent.lib.logs.ApacheLogParseMapOutputOperator;
-import org.apache.hadoop.conf.Configuration;
/**
* An implementation of Streaming Application that reads from a local apache log file on the fly and dumps the parsed data to output.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java
index d624e08..0eef6dc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java
@@ -16,16 +16,16 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
-import com.datatorrent.api.*;
-import com.datatorrent.api.DAG.Locality;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.logs.ApacheLogParseMapOutputOperator;
-import java.util.ArrayList;
-import java.util.List;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.logs.ApacheLogParseMapOutputOperator;
import org.apache.hadoop.conf.Configuration;
+import com.datatorrent.api.DAG;
+import com.datatorrent.api.DAG.Locality;
+import com.datatorrent.api.StreamingApplication;
+
/**
* An implementation of Streaming Application that generates apache log file on the fly and dumps the parsed data to output.
*
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java
index bd226e3..a58c39e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java
@@ -16,19 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
+import java.io.IOException;
import java.util.HashMap;
import java.util.Map;
-import java.io.IOException;
import javax.validation.constraints.NotNull;
-import com.datatorrent.lib.logs.InformationExtractor;
-import com.maxmind.geoip.Location;
-import com.maxmind.geoip.LookupService;
-
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.logs.InformationExtractor;
+
+import com.maxmind.geoip.Location;
+import com.maxmind.geoip.LookupService;
/**
* An implementation of InformationExtractor that extracts Geo information from an IP address using maxmind API .
@@ -67,8 +67,7 @@ public class GeoIPExtractor implements InformationExtractor
{
try {
reader = new LookupService(databasePath, LookupService.GEOIP_MEMORY_CACHE | LookupService.GEOIP_CHECK_CACHE);
- }
- catch (IOException ex) {
+ } catch (IOException ex) {
throw new RuntimeException(ex);
}
}
@@ -90,8 +89,7 @@ public class GeoIPExtractor implements InformationExtractor
m.put("ipRegion", location.region);
m.put("ipCity", location.city);
}
- }
- catch (Exception ex) {
+ } catch (Exception ex) {
LOG.error("Caught exception when looking up Geo IP for {}:", value, ex);
}
return m;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java
index bb87f03..1b6b4b6 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java
@@ -16,20 +16,20 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
-import java.util.Map;
-
-import com.datatorrent.lib.logs.InformationExtractor;
import java.text.DateFormat;
import java.text.ParseException;
import java.text.SimpleDateFormat;
import java.util.Date;
import java.util.HashMap;
+import java.util.Map;
import javax.validation.constraints.NotNull;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.logs.InformationExtractor;
+
/**
* An implementation of InformationExtractor that extracts the time stamp in milliseconds from epoch from an arbitrary date string.
* <p>
@@ -69,8 +69,7 @@ public class TimestampExtractor implements InformationExtractor
try {
Date date = dateFormat.parse((String)value);
m.put("timestamp", date.getTime());
- }
- catch (ParseException ex) {
+ } catch (ParseException ex) {
LOG.error("Error parsing \"{}\" to timestamp using \"{}\":", value, dateFormatString, ex);
}
return m;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java
index f73ff3d..799eb48 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java
@@ -16,20 +16,20 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
import java.util.HashMap;
import java.util.Map;
+import java.util.concurrent.TimeUnit;
+
+import org.apache.apex.malhar.lib.logs.InformationExtractor;
-import com.datatorrent.lib.logs.InformationExtractor;
import com.google.common.cache.Cache;
import com.google.common.cache.CacheBuilder;
-import java.util.concurrent.TimeUnit;
-
import net.sf.uadetector.ReadableUserAgent;
-import net.sf.uadetector.service.UADetectorServiceFactory;
import net.sf.uadetector.UserAgentStringParser;
+import net.sf.uadetector.service.UADetectorServiceFactory;
/**
* An implementation of Information Extractor that extracts the browser and the OS from a user-agent string.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java
index d99c7b5..67b151b 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java
@@ -20,4 +20,4 @@
* Apache log application, operators, and utilities.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java
index 738a146..e44c97a 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java
@@ -16,21 +16,21 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.apachelog.zmq;
+package org.apache.apex.malhar.contrib.apachelog.zmq;
+import org.apache.apex.malhar.contrib.zmq.SimpleSinglePortZeroMQPullStringInputOperator;
+import org.apache.apex.malhar.lib.algo.UniqueCounter;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.logs.ApacheLogParseOperator;
+import org.apache.apex.malhar.lib.math.Sum;
import org.apache.hadoop.conf.Configuration;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DAG;
import com.datatorrent.api.DAG.Locality;
-import com.datatorrent.api.annotation.ApplicationAnnotation;
import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.contrib.zmq.SimpleSinglePortZeroMQPullStringInputOperator;
-import com.datatorrent.lib.algo.UniqueCounter;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.logs.ApacheLogParseOperator;
-import com.datatorrent.lib.math.Sum;
+import com.datatorrent.api.annotation.ApplicationAnnotation;
/**
* <p>An implementation of Streaming Application that populates the DAG. </p>
@@ -40,7 +40,7 @@ import com.datatorrent.lib.math.Sum;
* @tag: streaming
* @since 0.3.2
*/
-@ApplicationAnnotation(name="ApacheLog")
+@ApplicationAnnotation(name = "ApacheLog")
public class Application implements StreamingApplication
{
private Locality locality = null;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java
index f863d41..52200fe 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
import java.io.IOException;
import java.io.InputStream;
@@ -24,6 +24,7 @@ import java.io.InputStream;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.io.fs.AbstractFileInputOperator;
import org.apache.apex.malhar.lib.wal.FSWindowDataManager;
import org.apache.avro.AvroRuntimeException;
import org.apache.avro.file.DataFileStream;
@@ -37,7 +38,6 @@ import com.google.common.annotations.VisibleForTesting;
import com.datatorrent.api.AutoMetric;
import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.lib.io.fs.AbstractFileInputOperator;
/**
* <p>
diff --git a/contrib/src/main/java/com/datatorrent/contrib/avro/AvroFileToPojoModule.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileToPojoModule.java
similarity index 98%
rename from contrib/src/main/java/com/datatorrent/contrib/avro/AvroFileToPojoModule.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileToPojoModule.java
index 8ad00df..4637b8f 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/avro/AvroFileToPojoModule.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileToPojoModule.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
import org.apache.apex.malhar.lib.wal.FSWindowDataManager;
import org.apache.avro.generic.GenericRecord;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java
index 012e977..176f5eb 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
import java.text.ParseException;
import java.util.List;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java
index 2acf98c..5539562 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
import java.util.ArrayList;
import java.util.List;
@@ -25,6 +25,9 @@ import java.util.StringTokenizer;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.FieldInfo.SupportType;
+import org.apache.apex.malhar.lib.util.PojoUtils;
import org.apache.avro.AvroRuntimeException;
import org.apache.avro.generic.GenericRecord;
import org.apache.commons.lang3.ClassUtils;
@@ -40,9 +43,6 @@ import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.FieldInfo.SupportType;
-import com.datatorrent.lib.util.PojoUtils;
/**
* <p>
@@ -87,7 +87,7 @@ public class AvroToPojo extends BaseOperator
@AutoMetric
@VisibleForTesting
int fieldErrorCount = 0;
-
+
public final transient DefaultOutputPort<GenericRecord> errorPort = new DefaultOutputPort<GenericRecord>();
/**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java
index 41c56e3..17a4ecc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
import java.io.IOException;
import java.util.ArrayList;
@@ -25,6 +25,8 @@ import java.util.List;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
import org.apache.avro.AvroRuntimeException;
import org.apache.avro.Schema;
import org.apache.avro.Schema.Field;
@@ -42,8 +44,6 @@ import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.annotation.InputPortFieldAnnotation;
import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
/**
* <p>
@@ -63,7 +63,6 @@ import com.datatorrent.lib.util.PojoUtils.Getter;
@InterfaceStability.Evolving
public class PojoToAvro extends BaseOperator
{
-
private List<Field> columnNames;
private Class<?> cls;
@@ -155,7 +154,7 @@ public class PojoToAvro extends BaseOperator
* @return Getter
*/
private Getter<?, ?> generateGettersForField(Class<?> cls, String inputFieldName)
- throws NoSuchFieldException, SecurityException
+ throws NoSuchFieldException, SecurityException
{
java.lang.reflect.Field f = cls.getDeclaredField(inputFieldName);
Class<?> c = ClassUtils.primitiveToWrapper(f.getType());
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java
index 7bd47fc..3153387 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java
@@ -16,18 +16,18 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
import com.datastax.driver.core.PagingState;
import com.datastax.driver.core.ResultSet;
import com.datastax.driver.core.Row;
import com.datastax.driver.core.SimpleStatement;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.netlet.util.DTThrowable;
@@ -43,8 +43,8 @@ import com.datatorrent.netlet.util.DTThrowable;
* @tags cassandra
* @since 1.0.2
*/
-public abstract class AbstractCassandraInputOperator<T> extends AbstractStoreInputOperator<T, CassandraStore> {
-
+public abstract class AbstractCassandraInputOperator<T> extends AbstractStoreInputOperator<T, CassandraStore>
+{
private static final Logger logger = LoggerFactory.getLogger(AbstractCassandraInputOperator.class);
private PagingState nextPageState;
private int fetchSize;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java
index 9048383..f0f590d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java
@@ -16,12 +16,14 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import java.util.Collection;
import javax.annotation.Nonnull;
+import org.apache.apex.malhar.lib.db.AbstractBatchTransactionableStoreOutputOperator;
+
import com.datastax.driver.core.BatchStatement;
import com.datastax.driver.core.PreparedStatement;
import com.datastax.driver.core.Statement;
@@ -29,7 +31,6 @@ import com.datastax.driver.core.exceptions.DriverException;
import com.datatorrent.api.Context;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.Operator.ActivationListener;
-import com.datatorrent.lib.db.AbstractBatchTransactionableStoreOutputOperator;
/**
* <p>
@@ -86,8 +87,7 @@ public abstract class AbstractCassandraTransactionableOutputOperator<T> extends
public void processBatch(Collection<T> tuples)
{
BatchStatement batchCommand = store.getBatchCommand();
- for(T tuple: tuples)
- {
+ for (T tuple: tuples) {
batchCommand.add(setStatementParameters(updateCommand, tuple));
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java
index f287385..cadf157 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import java.io.IOException;
import java.io.InputStream;
@@ -949,7 +949,7 @@ public abstract class AbstractUpsertOutputOperator extends BaseOperator implemen
/**
* Implementing concrete Operator instances define the Connection Builder properties by implementing this method
- * Please refer to {@link com.datatorrent.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} for
+ * Please refer to {@link org.apache.apex.malhar.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} for
* an example implementation of the ConnectionStateManager instantiation.
* Note that if this method is returning null, the connection properties are
* fetched from a properties file loaded from the classpath.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java
index f43777f..cab83e8 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java
@@ -16,17 +16,30 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import java.math.BigDecimal;
-import java.util.*;
+import java.util.ArrayList;
+import java.util.Date;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.UUID;
import javax.validation.constraints.Min;
import javax.validation.constraints.NotNull;
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Setter;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterBoolean;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterDouble;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterFloat;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterInt;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterLong;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
import com.datastax.driver.core.BoundStatement;
import com.datastax.driver.core.ColumnDefinitions;
@@ -39,9 +52,6 @@ import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.Operator;
import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.*;
/**
* <p>
@@ -287,8 +297,7 @@ public class CassandraPOJOInputOperator extends AbstractCassandraInputOperator<O
try {
// This code will be replaced after integration of creating POJOs on the fly utility.
obj = pojoClass.newInstance();
- }
- catch (InstantiationException | IllegalAccessException ex) {
+ } catch (InstantiationException | IllegalAccessException ex) {
throw new RuntimeException(ex);
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java
index a191bb0..5f7d52a 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java
@@ -16,28 +16,44 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import java.lang.reflect.Field;
import java.math.BigDecimal;
-import java.util.*;
+import java.util.ArrayList;
+import java.util.Date;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.UUID;
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.google.common.collect.Lists;
-import com.datastax.driver.core.*;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterBoolean;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterDouble;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterFloat;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterInt;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterLong;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
+
+import com.datastax.driver.core.BoundStatement;
+import com.datastax.driver.core.ColumnDefinitions;
+import com.datastax.driver.core.DataType;
+import com.datastax.driver.core.LocalDate;
+import com.datastax.driver.core.PreparedStatement;
+import com.datastax.driver.core.Statement;
import com.datastax.driver.core.exceptions.DriverException;
+import com.google.common.collect.Lists;
import com.datatorrent.api.AutoMetric;
import com.datatorrent.api.Context;
import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.annotation.InputPortFieldAnnotation;
import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.*;
/**
* <p>
@@ -106,10 +122,11 @@ public class CassandraPOJOOutputOperator extends AbstractCassandraTransactionabl
@Override
public void activate(Context.OperatorContext context)
{
- com.datastax.driver.core.ResultSet rs = store.getSession().execute("select * from " + store.keyspace + "." + tablename);
+ com.datastax.driver.core.ResultSet rs
+ = store.getSession().execute("select * from " + store.keyspace + "." + tablename);
final ColumnDefinitions rsMetaData = rs.getColumnDefinitions();
- if(fieldInfos == null) {
+ if (fieldInfos == null) {
populateFieldInfosFromPojo(rsMetaData);
}
@@ -228,13 +245,12 @@ public class CassandraPOJOOutputOperator extends AbstractCassandraTransactionabl
}
}
String statement
- = "INSERT INTO " + store.keyspace + "."
- + tablename
- + " (" + queryfields.toString() + ") "
- + "VALUES (" + values.toString() + ");";
+ = "INSERT INTO " + store.keyspace + "."
+ + tablename
+ + " (" + queryfields.toString() + ") "
+ + "VALUES (" + values.toString() + ");";
LOG.debug("statement is {}", statement);
return store.getSession().prepare(statement);
-
}
@Override
@@ -316,6 +332,7 @@ public class CassandraPOJOOutputOperator extends AbstractCassandraTransactionabl
errorRecords++;
}
}
+
/**
* A list of {@link FieldInfo}s where each item maps a column name to a pojo field name.
*/
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java
index 8b435c9..179ac1a 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import java.math.BigDecimal;
import java.util.Date;
@@ -25,11 +25,12 @@ import java.util.Map;
import java.util.Set;
import java.util.UUID;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+
import com.datastax.driver.core.BoundStatement;
import com.datastax.driver.core.DataType;
import com.datastax.driver.core.LocalDate;
import com.datastax.driver.core.TypeCodec;
-import com.datatorrent.lib.util.PojoUtils;
/**
* Used to manage simple data type based getters for given cassandra columns
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java
index 713179c..a797bab 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import java.util.EnumSet;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java
index 5d9178c..0c298aa 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java
@@ -16,19 +16,20 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import javax.validation.constraints.NotNull;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.Connectable;
+
import com.datastax.driver.core.Cluster;
import com.datastax.driver.core.ProtocolVersion;
import com.datastax.driver.core.Session;
import com.datastax.driver.core.exceptions.DriverException;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.Connectable;
/**
* A {@link Connectable} that uses cassandra to connect to stores and implements Connectable interface.
@@ -67,7 +68,8 @@ public class CassandraStore implements Connectable
*
* @param keyspace keyspace.
*/
- public void setKeyspace(String keyspace) {
+ public void setKeyspace(String keyspace)
+ {
this.keyspace = keyspace;
}
@@ -106,7 +108,8 @@ public class CassandraStore implements Connectable
}
@NotNull
- public String getNode() {
+ public String getNode()
+ {
return node;
}
@@ -115,15 +118,18 @@ public class CassandraStore implements Connectable
*
* @param node node
*/
- public void setNode(@NotNull String node) {
+ public void setNode(@NotNull String node)
+ {
this.node = node;
}
- public Cluster getCluster() {
+ public Cluster getCluster()
+ {
return cluster;
}
- public Session getSession() {
+ public Session getSession()
+ {
return session;
}
@@ -167,20 +173,18 @@ public class CassandraStore implements Connectable
public void connect()
{
try {
- if(cluster==null)
+ if (cluster == null) {
buildCluster();
+ }
session = cluster.connect();
logger.debug("Cassandra connection Success");
- }
- catch (DriverException ex) {
+ } catch (DriverException ex) {
throw new RuntimeException("closing database resource", ex);
- }
- catch (Throwable t) {
+ } catch (Throwable t) {
DTThrowable.rethrow(t);
}
}
-
/**
* Close connection.
*/
@@ -190,11 +194,9 @@ public class CassandraStore implements Connectable
try {
session.close();
cluster.close();
- }
- catch (DriverException ex) {
+ } catch (DriverException ex) {
throw new RuntimeException("closing database resource", ex);
- }
- catch (Throwable t) {
+ } catch (Throwable t) {
DTThrowable.rethrow(t);
}
}
@@ -204,8 +206,7 @@ public class CassandraStore implements Connectable
{
try {
return !session.isClosed();
- }
- catch (DriverException ex) {
+ } catch (DriverException ex) {
throw new RuntimeException("closing database resource", ex);
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java
index 4c5a71e..29b6ce9 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java
@@ -16,14 +16,18 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import javax.annotation.Nonnull;
-import com.datastax.driver.core.*;
-import com.datastax.driver.core.exceptions.DriverException;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
-import com.datatorrent.lib.db.TransactionableStore;
+import com.datastax.driver.core.BatchStatement;
+import com.datastax.driver.core.BoundStatement;
+import com.datastax.driver.core.PreparedStatement;
+import com.datastax.driver.core.ResultSet;
+import com.datastax.driver.core.Statement;
+import com.datastax.driver.core.exceptions.DriverException;
/**
* <p>Provides transaction support to the operators by implementing TransactionableStore abstract methods. </p>
@@ -34,8 +38,8 @@ import com.datatorrent.lib.db.TransactionableStore;
* @tags cassandra, transactional
* @since 1.0.2
*/
-public class CassandraTransactionalStore extends CassandraStore implements TransactionableStore {
-
+public class CassandraTransactionalStore extends CassandraStore implements TransactionableStore
+{
public static String DEFAULT_APP_ID_COL = "dt_app_id";
public static String DEFAULT_OPERATOR_ID_COL = "dt_operator_id";
public static String DEFAULT_WINDOW_COL = "dt_window";
@@ -116,7 +120,8 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
this.metaTableWindowColumn = windowColumn;
}
- public Statement getLastWindowUpdateStatement() {
+ public Statement getLastWindowUpdateStatement()
+ {
return lastWindowUpdateStatement;
}
@@ -130,26 +135,24 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
{
super.connect();
try {
- String command = "SELECT " + metaTableWindowColumn + " FROM " + keyspace +"."+ metaTable + " WHERE " + metaTableAppIdColumn +
+ String command = "SELECT " + metaTableWindowColumn + " FROM " + keyspace + "." + metaTable +
+ " WHERE " + metaTableAppIdColumn +
" = ? AND " + metaTableOperatorIdColumn + " = ?";
logger.debug(command);
lastWindowFetchCommand = session.prepare(command);
- command = "UPDATE " + keyspace +"."+ metaTable + " SET " + metaTableWindowColumn + " = ? where " + metaTableAppIdColumn + " = ? " +
+ command = "UPDATE " + keyspace + "." + metaTable + " SET " + metaTableWindowColumn + " = ? where " + metaTableAppIdColumn + " = ? " +
" and " + metaTableOperatorIdColumn + " = ?";
logger.debug(command);
lastWindowUpdateCommand = session.prepare(command);
- command = "DELETE FROM " + keyspace +"."+ metaTable + " where " + metaTableAppIdColumn + " = ? and " +
+ command = "DELETE FROM " + keyspace + "." + metaTable + " where " + metaTableAppIdColumn + " = ? and " +
metaTableOperatorIdColumn + " = ?";
logger.debug(command);
lastWindowDeleteCommand = session.prepare(command);
- }
- catch (DriverException e) {
+ } catch (DriverException e) {
throw new RuntimeException(e);
- }
- catch (Exception e)
- {
+ } catch (Exception e) {
throw new RuntimeException(e);
}
}
@@ -160,8 +163,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
if (lastWindowUpdateCommand != null) {
try {
lastWindowUpdateCommand.disableTracing();
- }
- catch (DriverException e) {
+ } catch (DriverException e) {
throw new RuntimeException(e);
}
}
@@ -208,8 +210,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
}
lastWindowFetchCommand.disableTracing();
return lastWindow;
- }
- catch (DriverException ex) {
+ } catch (DriverException ex) {
throw new RuntimeException(ex);
}
}
@@ -221,8 +222,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
BoundStatement boundStatement = new BoundStatement(lastWindowUpdateCommand);
lastWindowUpdateStatement = boundStatement.bind(windowId,appId,operatorId);
batchCommand.add(lastWindowUpdateStatement);
- }
- catch (DriverException e) {
+ } catch (DriverException e) {
throw new RuntimeException(e);
}
}
@@ -235,8 +235,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
lastWindowDeleteStatement = boundStatement.bind(appId,operatorId);
session.execute(lastWindowDeleteStatement);
lastWindowDeleteCommand.disableTracing();
- }
- catch (DriverException e) {
+ } catch (DriverException e) {
throw new RuntimeException(e);
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java
index f089137..0eccb22 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import java.io.Serializable;
import java.net.InetAddress;
@@ -64,7 +64,7 @@ import static com.google.common.base.Preconditions.checkNotNull;
* </p>
* <p> A typical implementation of the ConnectionBuilder would like this:
* <pre>
- *
+ *
* public ConnectionStateManager.ConnectionBuilder withConnectionBuilder()
* {
* return ConnectionStateManager.withNewBuilder()
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java
index 92be546..10ffb08 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
import com.datastax.driver.core.ConsistencyLevel;
@@ -122,7 +122,7 @@ public class UpsertExecutionContext<T>
/**
* This decides if we want to override the default TTL if at all set in the
- * {@link com.datatorrent.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} that is used to execute a
+ * {@link org.apache.apex.malhar.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} that is used to execute a
* mutation. Note that TTLs are not mandatory for mutations.
* Also it is supported to have TTLs only for the current execution context but not set a default at the
* connection state manager level
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java
index 1baae5e..93c5518 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java
@@ -17,4 +17,4 @@
* under the License.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java
index 1cd4eb5..3c53a49 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
import java.io.IOException;
import java.util.Collection;
@@ -24,21 +24,20 @@ import java.util.List;
import java.util.Map;
import java.util.concurrent.TimeUnit;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
+import org.apache.apex.malhar.lib.util.KryoCloneUtils;
+
import com.couchbase.client.CouchbaseClient;
import com.couchbase.client.vbucket.config.Config;
import com.google.common.annotations.VisibleForTesting;
import com.google.common.collect.Lists;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
-
import com.datatorrent.api.Context;
import com.datatorrent.api.DefaultPartition;
import com.datatorrent.api.Partitioner;
-
-import com.datatorrent.lib.util.KryoCloneUtils;
import com.datatorrent.netlet.util.DTThrowable;
/**
@@ -94,8 +93,7 @@ public abstract class AbstractCouchBaseInputOperator<T> extends AbstractStoreInp
}
try {
clientPartition = store.connectServer(serverURIString);
- }
- catch (IOException ex) {
+ } catch (IOException ex) {
DTThrowable.rethrow(ex);
}
}
@@ -116,16 +114,16 @@ public abstract class AbstractCouchBaseInputOperator<T> extends AbstractStoreInp
List<String> keys = getKeys();
Object result = null;
for (String key: keys) {
- int master = conf.getMaster(conf.getVbucketByKey(key));
- if (master == getServerIndex()) {
- result = clientPartition.get(key);
- }
+ int master = conf.getMaster(conf.getVbucketByKey(key));
+ if (master == getServerIndex()) {
+ result = clientPartition.get(key);
}
+ }
- if (result != null) {
- T tuple = getTuple(result);
- outputPort.emit(tuple);
- }
+ if (result != null) {
+ T tuple = getTuple(result);
+ outputPort.emit(tuple);
+ }
}
public abstract T getTuple(Object object);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java
index ef3be01..8cc92f2 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java
@@ -16,23 +16,24 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
import java.util.HashMap;
import java.util.Iterator;
import java.util.TreeMap;
-import net.spy.memcached.internal.OperationCompletionListener;
-import net.spy.memcached.internal.OperationFuture;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.datatorrent.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
+import org.apache.apex.malhar.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
import com.datatorrent.api.Context.OperatorContext;
-
import com.datatorrent.netlet.util.DTThrowable;
+import net.spy.memcached.internal.OperationCompletionListener;
+import net.spy.memcached.internal.OperationFuture;
+
+
/**
* AbstractCouchBaseOutputOperator which extends Transactionable Store Output Operator.
* Classes extending from this operator should implement the abstract functionality of generateKey, getValue and insertOrUpdate.
@@ -108,10 +109,10 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
id++;
String key = getKey(tuple);
Object value = getValue(tuple);
- if(!(value instanceof Boolean) && !(value instanceof Integer) && !(value instanceof String) && !(value instanceof Float) && !(value instanceof Double) && !(value instanceof Character) && !(value instanceof Long) && !(value instanceof Short) && !(value instanceof Byte)){
- if (serializer != null) {
- value = serializer.serialize(value);
- }
+ if (!(value instanceof Boolean) && !(value instanceof Integer) && !(value instanceof String) && !(value instanceof Float) && !(value instanceof Double) && !(value instanceof Character) && !(value instanceof Long) && !(value instanceof Short) && !(value instanceof Byte)) {
+ if (serializer != null) {
+ value = serializer.serialize(value);
+ }
}
OperationFuture<Boolean> future = processKeyValue(key, value);
synchronized (syncObj) {
@@ -135,7 +136,7 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
public void waitForQueueSize(int sizeOfQueue)
{
long startTms = System.currentTimeMillis();
- long elapsedTime ;
+ long elapsedTime;
while (numTuples > sizeOfQueue) {
synchronized (syncObj) {
if (numTuples > sizeOfQueue) {
@@ -143,12 +144,10 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
elapsedTime = System.currentTimeMillis() - startTms;
if (elapsedTime >= store.timeout) {
throw new RuntimeException("Timed out waiting for space in queue");
- }
- else {
+ } else {
syncObj.wait(store.timeout - elapsedTime);
}
- }
- catch (InterruptedException ex) {
+ } catch (InterruptedException ex) {
DTThrowable.rethrow(ex);
}
}
@@ -167,7 +166,7 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
public void onComplete(OperationFuture<?> f) throws Exception
{
if (!((Boolean)f.get())) {
- logger.error("Operation failed {}" , f);
+ logger.error("Operation failed {}", f);
failure = true;
return;
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java
index f33d804..0abe06f 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
import net.spy.memcached.internal.OperationFuture;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java
index 188d989..0f968bd 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java
@@ -16,11 +16,11 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
-import com.datatorrent.netlet.util.DTThrowable;
import java.io.IOException;
import org.codehaus.jackson.map.ObjectMapper;
+import com.datatorrent.netlet.util.DTThrowable;
/**
* <p>CouchBaseJSONSerializer</p>
@@ -43,8 +43,7 @@ public class CouchBaseJSONSerializer implements CouchBaseSerializer
String value = null;
try {
value = mapper.writeValueAsString(o);
- }
- catch (IOException ex) {
+ } catch (IOException ex) {
DTThrowable.rethrow(ex);
}
return value;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java
index 2378aec..91e3b17 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
import java.io.IOException;
import java.util.Iterator;
@@ -24,12 +24,16 @@ import java.util.Iterator;
import javax.validation.constraints.Min;
import javax.validation.constraints.NotNull;
-import com.couchbase.client.protocol.views.*;
-
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.codehaus.jackson.map.ObjectMapper;
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
+
+import com.couchbase.client.protocol.views.Query;
+import com.couchbase.client.protocol.views.Stale;
+import com.couchbase.client.protocol.views.View;
+import com.couchbase.client.protocol.views.ViewResponse;
+import com.couchbase.client.protocol.views.ViewRow;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
import com.datatorrent.api.Context.OperatorContext;
/**
@@ -167,8 +171,7 @@ public class CouchBasePOJOInputOperator extends AbstractStoreInputOperator<Objec
super.setup(context);
try {
className = Class.forName(outputClass);
- }
- catch (ClassNotFoundException ex) {
+ } catch (ClassNotFoundException ex) {
throw new RuntimeException(ex);
}
view = store.getInstance().getView(designDocumentName, viewName);
@@ -198,8 +201,7 @@ public class CouchBasePOJOInputOperator extends AbstractStoreInputOperator<Objec
Object outputObj = null;
try {
outputObj = objectMapper.readValue(document.toString(), className);
- }
- catch (IOException ex) {
+ } catch (IOException ex) {
throw new RuntimeException(ex);
}
outputPort.emit(outputObj);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java
index f15758e..0769e82 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
/**
*
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java
index 238047e..fb5ba1d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
import java.io.IOException;
import java.net.URI;
@@ -28,6 +28,10 @@ import java.util.concurrent.TimeUnit;
import javax.annotation.Nonnull;
import javax.validation.constraints.Min;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.Connectable;
+
import com.couchbase.client.CouchbaseClient;
import com.couchbase.client.CouchbaseConnectionFactoryBuilder;
import com.couchbase.client.vbucket.ConfigurationProvider;
@@ -35,11 +39,6 @@ import com.couchbase.client.vbucket.ConfigurationProviderHTTP;
import com.couchbase.client.vbucket.config.Bucket;
import com.couchbase.client.vbucket.config.Config;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import com.datatorrent.lib.db.Connectable;
-
import com.datatorrent.netlet.util.DTThrowable;
/**
@@ -108,6 +107,7 @@ public class CouchBaseStore implements Connectable
{
this.queueSize = queueSize;
}
+
protected Integer maxTuples = 1000;
protected int blockTime = 1000;
protected long timeout = 10000;
@@ -192,8 +192,7 @@ public class CouchBaseStore implements Connectable
{
try {
connect();
- }
- catch (IOException ex) {
+ } catch (IOException ex) {
DTThrowable.rethrow(ex);
}
ConfigurationProvider configurationProvider = new ConfigurationProviderHTTP(baseURIs, userConfig, passwordConfig);
@@ -201,8 +200,7 @@ public class CouchBaseStore implements Connectable
Config conf = configBucket.getConfig();
try {
disconnect();
- }
- catch (IOException ex) {
+ } catch (IOException ex) {
DTThrowable.rethrow(ex);
}
return conf;
@@ -216,8 +214,7 @@ public class CouchBaseStore implements Connectable
for (String url : tokens) {
try {
uri = new URI("http", url, "/pools", null, null);
- }
- catch (URISyntaxException ex) {
+ } catch (URISyntaxException ex) {
DTThrowable.rethrow(ex);
}
baseURIs.add(uri);
@@ -228,9 +225,8 @@ public class CouchBaseStore implements Connectable
cfb.setOpQueueMaxBlockTime(blockTime); // wait up to 10 second when trying to enqueue an operation
client = new CouchbaseClient(cfb.buildCouchbaseConnection(baseURIs, bucket, password));
//client = new CouchbaseClient(baseURIs, "default", "");
- }
- catch (IOException e) {
- logger.error("Error connecting to Couchbase:" , e);
+ } catch (IOException e) {
+ logger.error("Error connecting to Couchbase:", e);
DTThrowable.rethrow(e);
}
}
@@ -241,15 +237,13 @@ public class CouchBaseStore implements Connectable
CouchbaseClient clientPartition = null;
try {
nodes.add(new URI("http",urlString,"/pools", null, null));
- }
- catch (URISyntaxException ex) {
+ } catch (URISyntaxException ex) {
DTThrowable.rethrow(ex);
}
try {
clientPartition = new CouchbaseClient(nodes, bucket, password);
- }
- catch (IOException e) {
- logger.error("Error connecting to Couchbase:" , e);
+ } catch (IOException e) {
+ logger.error("Error connecting to Couchbase:", e);
DTThrowable.rethrow(e);
}
return clientPartition;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java
index 0c57ca6..2f0ce59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java
@@ -16,19 +16,23 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
-import java.io.*;
+import java.io.ByteArrayInputStream;
+import java.io.ByteArrayOutputStream;
+import java.io.DataInputStream;
+import java.io.DataOutputStream;
+import java.io.IOException;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.TimeUnit;
-import com.couchbase.client.CouchbaseClient;
-import com.couchbase.client.CouchbaseConnectionFactoryBuilder;
-
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.datatorrent.lib.db.TransactionableStore;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
+
+import com.couchbase.client.CouchbaseClient;
+import com.couchbase.client.CouchbaseConnectionFactoryBuilder;
import com.datatorrent.netlet.util.DTThrowable;
@@ -92,9 +96,8 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
cfb.setOpTimeout(timeout); // wait up to 10 seconds for an operation to succeed
cfb.setOpQueueMaxBlockTime(blockTime); // wait up to 10 second when trying to enqueue an operation
clientMeta = new CouchbaseClient(cfb.buildCouchbaseConnection(baseURIs, bucketMeta, passwordMeta));
- }
- catch (IOException e) {
- logger.error("Error connecting to Couchbase: " , e);
+ } catch (IOException e) {
+ logger.error("Error connecting to Couchbase: ", e);
DTThrowable.rethrow(e);
}
}
@@ -119,11 +122,9 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
String key = appId + "_" + operatorId + "_" + lastWindowValue;
try {
clientMeta.set(key, WindowIdBytes).get();
- }
- catch (InterruptedException ex) {
+ } catch (InterruptedException ex) {
DTThrowable.rethrow(ex);
- }
- catch (ExecutionException ex) {
+ } catch (ExecutionException ex) {
DTThrowable.rethrow(ex);
}
@@ -163,8 +164,7 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
try {
result = dos.readLong();
dos.close();
- }
- catch (IOException e) {
+ } catch (IOException e) {
DTThrowable.rethrow(e);
}
return result;
@@ -179,8 +179,7 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
dos.writeLong(l);
result = baos.toByteArray();
dos.close();
- }
- catch (IOException e) {
+ } catch (IOException e) {
DTThrowable.rethrow(e);
}
return result;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java
index 5221b02..64ce9e2 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java
@@ -16,18 +16,18 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
import java.util.ArrayList;
import javax.validation.constraints.NotNull;
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
/**
* An implementation of Couchbase Output Operator which takes a POJO,serializes it into key,value
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java
index 1a6d05b..7b7dc59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java
@@ -17,4 +17,4 @@
* under the License.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java
index 1935e5c..ad3e797 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java
@@ -16,20 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
import java.io.IOException;
import java.util.List;
import javax.validation.constraints.Min;
-import com.google.common.base.Throwables;
-
import org.ektorp.ViewQuery;
import org.ektorp.ViewResult;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
+import com.google.common.base.Throwables;
import com.datatorrent.api.Context;
@@ -92,8 +91,7 @@ public abstract class AbstractCouchDBInputOperator<T> extends AbstractStoreInput
T tuple = getTuple(row);
outputPort.emit(tuple);
}
- }
- catch (Throwable cause) {
+ } catch (Throwable cause) {
Throwables.propagate(cause);
}
if (rows.size() > 0) {
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java
index 5e6c24b..8080eda 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java
@@ -16,10 +16,9 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
-import java.util.Map;
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
/**
* Generic base output adaptor which saves tuples in the CouchDb. Subclasses should provide implementation for getting Document Id. <br/>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java
index 94c814f..e39a91c 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java
@@ -16,15 +16,15 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchdb;
-
-import org.codehaus.jackson.map.ObjectMapper;
-import org.ektorp.ViewResult;
+package org.apache.apex.malhar.contrib.couchdb;
import java.io.IOException;
import java.util.HashMap;
import java.util.Map;
+import org.codehaus.jackson.map.ObjectMapper;
+import org.ektorp.ViewResult;
+
/**
* A base CouchDb input adaptor that emits a map. <br/>
* <p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java
index 8111727..6d5defa 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
import java.io.IOException;
import java.util.ArrayList;
@@ -24,14 +24,18 @@ import java.util.List;
import javax.validation.constraints.NotNull;
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.codehaus.jackson.JsonNode;
import org.codehaus.jackson.map.ObjectMapper;
import org.ektorp.ViewQuery;
import org.ektorp.ViewResult.Row;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Setter;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterBoolean;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterDouble;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterInt;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterLong;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.*;
import com.datatorrent.api.Context.OperatorContext;
/**
@@ -184,8 +188,7 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
try {
// This code will be replaced after integration of creating POJOs on the fly utility.
objectClass = Class.forName(outputClass);
- }
- catch (ClassNotFoundException ex) {
+ } catch (ClassNotFoundException ex) {
throw new RuntimeException(ex);
}
@@ -197,18 +200,15 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
Class<?> type = null;
try {
type = objectClass.getDeclaredField(columns.get(i)).getType();
- }
- catch (NoSuchFieldException ex) {
+ } catch (NoSuchFieldException ex) {
throw new RuntimeException(ex);
- }
- catch (SecurityException ex) {
+ } catch (SecurityException ex) {
throw new RuntimeException(ex);
}
fieldType.add(type);
if (type.isPrimitive()) {
setterDoc.add(PojoUtils.constructSetter(objectClass, expressions.get(i), type));
- }
- else {
+ } else {
setterDoc.add(PojoUtils.createSetter(objectClass, expressions.get(i), type));
}
}
@@ -222,11 +222,9 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
Object obj;
try {
obj = objectClass.newInstance();
- }
- catch (InstantiationException ex) {
+ } catch (InstantiationException ex) {
throw new RuntimeException(ex);
- }
- catch (IllegalAccessException ex) {
+ } catch (IllegalAccessException ex) {
throw new RuntimeException(ex);
}
@@ -240,21 +238,16 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
if (type.isPrimitive()) {
if (type == int.class) {
((SetterInt)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getIntValue());
- }
- else if (type == boolean.class) {
+ } else if (type == boolean.class) {
((SetterBoolean)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getBooleanValue());
- }
- else if (type == long.class) {
+ } else if (type == long.class) {
((SetterLong)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getLongValue());
- }
- else if (type == double.class) {
+ } else if (type == double.class) {
((SetterDouble)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getDoubleValue());
- }
- else {
+ } else {
throw new RuntimeException("Type is not supported");
}
- }
- else {
+ } else {
((Setter<Object, Object>)setterDoc.get(i)).set(obj, mapper.readValue(val.get(columns.get(i)), type));
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java
index f75473c..9d75272 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java
@@ -16,15 +16,14 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
import javax.validation.constraints.NotNull;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
import org.apache.hadoop.classification.InterfaceStability.Evolving;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
-
/**
* Implementation of {@link AbstractCouchDBOutputOperator} that saves a POJO in the couch database. <br/>
* <p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java
index 88c2eab..31e5423 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
import java.io.IOException;
import java.net.MalformedURLException;
@@ -32,7 +32,7 @@ import org.ektorp.http.HttpClient;
import org.ektorp.http.StdHttpClient;
import org.ektorp.impl.StdCouchDbInstance;
-import com.datatorrent.lib.db.Connectable;
+import org.apache.apex.malhar.lib.db.Connectable;
/**
* Implements a CouchDb store. <br/>
@@ -141,8 +141,7 @@ public class CouchDbStore implements Connectable
{
if (docId != null && dbConnector.contains(docId)) {
dbConnector.update(document);
- }
- else {
+ } else {
//create a document & if docId is null then couch db will generate a random id.
dbConnector.create(document);
}
@@ -166,8 +165,7 @@ public class CouchDbStore implements Connectable
if (dbUrl != null) {
try {
builder.url(dbUrl);
- }
- catch (MalformedURLException e) {
+ } catch (MalformedURLException e) {
throw new IllegalArgumentException(e.getMessage());
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java
index f83f93c..6c8b813 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
import java.util.Map;
@@ -34,7 +34,7 @@ public class MapBasedCouchDbOutputOperator extends AbstractCouchDBOutputOperator
@Override
public String getDocumentId(Map<Object, Object> tuple)
{
- return (String) tuple.get("_id");
+ return (String)tuple.get("_id");
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java
index be6ec34..13b7e17 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java
@@ -17,4 +17,4 @@
* under the License.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java
index 022530b..5d9e915 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java
@@ -16,14 +16,15 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
import org.elasticsearch.action.search.SearchRequestBuilder;
import org.elasticsearch.action.search.SearchResponse;
import org.elasticsearch.search.SearchHit;
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
+
import com.datatorrent.api.Context.OperatorContext;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
/**
* This is the base implementation for a non transactional input operator for ElasticSearch
@@ -53,7 +54,7 @@ public abstract class AbstractElasticSearchInputOperator<T, S extends ElasticSea
/**
* Initializing transient fields such as ElasticSearchConnectable, SearchRequestBuilder
*
- * @see com.datatorrent.lib.db.AbstractStoreInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
+ * @see org.apache.apex.malhar.lib.db.AbstractStoreInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
*/
@Override
public void setup(OperatorContext t1)
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java
index 0282ae8..34b4440 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
import java.util.Queue;
import java.util.concurrent.ArrayBlockingQueue;
@@ -27,9 +27,10 @@ import org.elasticsearch.action.bulk.BulkRequestBuilder;
import org.elasticsearch.action.bulk.BulkResponse;
import org.elasticsearch.action.index.IndexRequestBuilder;
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
+
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
/**
* This is the base implementation for a non-transactional batch output operator for ElasticSearch.
@@ -70,7 +71,7 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
/**
* Initialize transient fields such as {@code tupleBatch}
*
- * @see com.datatorrent.lib.db.AbstractStoreOutputOperator#setup(com.datatorrent.api.Context.OperatorContext)
+ * @see org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator#setup(com.datatorrent.api.Context.OperatorContext)
*/
@Override
public void setup(OperatorContext context)
@@ -83,7 +84,7 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
* Adds tuple to the queue.
* Calls {@link #processBatch()} if queue is full
*
- * @see com.datatorrent.lib.db.AbstractStoreOutputOperator#processTuple(java.lang.Object)
+ * @see org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator#processTuple(java.lang.Object)
*/
public void processTuple(T tuple)
{
@@ -128,7 +129,8 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
* @param tuple
* @return
*/
- protected IndexRequestBuilder getIndexRequestBuilder(T tuple){
+ protected IndexRequestBuilder getIndexRequestBuilder(T tuple)
+ {
IndexRequestBuilder indexRequestBuilder = new IndexRequestBuilder(store.client, getIndexName(tuple));
String id = getId(tuple);
if (id != null) {
@@ -176,22 +178,22 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
* @param tuple
* @return
*/
- protected abstract String getType(T tuple);
+ protected abstract String getType(T tuple);
/**
* @return the batchSize
*/
- public int getBatchSize()
- {
- return batchSize;
- }
+ public int getBatchSize()
+ {
+ return batchSize;
+ }
/**
* @param batchSize the batchSize to set
*/
- public void setBatchSize(int batchSize)
- {
- this.batchSize = batchSize;
- }
+ public void setBatchSize(int batchSize)
+ {
+ this.batchSize = batchSize;
+ }
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java
index 34eca95..c607393 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
import java.io.IOException;
@@ -25,7 +25,7 @@ import javax.validation.constraints.NotNull;
import org.elasticsearch.client.transport.TransportClient;
import org.elasticsearch.common.transport.InetSocketTransportAddress;
-import com.datatorrent.lib.db.Connectable;
+import org.apache.apex.malhar.lib.db.Connectable;
/**
* Elastic search base connector which has basic information for an operator <br>
@@ -91,7 +91,7 @@ public class ElasticSearchConnectable implements Connectable
/*
* (non-Javadoc)
*
- * @see com.datatorrent.lib.db.Connectable#connect()
+ * @see org.apache.apex.malhar.lib.db.Connectable#connect()
*/
@Override
public void connect() throws IOException
@@ -103,7 +103,7 @@ public class ElasticSearchConnectable implements Connectable
/*
* (non-Javadoc)
*
- * @see com.datatorrent.lib.db.Connectable#disconnect()
+ * @see org.apache.apex.malhar.lib.db.Connectable#disconnect()
*/
@Override
public void disconnect() throws IOException
@@ -116,7 +116,7 @@ public class ElasticSearchConnectable implements Connectable
/*
* (non-Javadoc)
*
- * @see com.datatorrent.lib.db.Connectable#isConnected()
+ * @see org.apache.apex.malhar.lib.db.Connectable#isConnected()
*/
@Override
public boolean isConnected()
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java
index dcbee9d..6774f10 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
import java.util.Map;
@@ -50,7 +50,7 @@ public abstract class ElasticSearchMapInputOperator<T extends Map<String, Object
/**
* {@link SearchRequestBuilder} properties which do not change for each window are set during operator initialization.
*
- * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
+ * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
*/
@Override
public void setup(OperatorContext t1)
@@ -63,14 +63,14 @@ public abstract class ElasticSearchMapInputOperator<T extends Map<String, Object
* (non-Javadoc)
*
* @see
- * com.datatorrent.contrib.elasticsearch.AbstractElasticSearchInputOperator#convertToTuple(org.elasticsearch.search
+ * org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchInputOperator#convertToTuple(org.elasticsearch.search
* .SearchHit)
*/
@Override
protected T convertToTuple(SearchHit hit)
{
Map<String, Object> tuple = hit.getSource();
- return (T) tuple;
+ return (T)tuple;
}
/**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java
index 8616938..d85f0a3 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
import java.util.Map;
@@ -50,7 +50,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
* (non-Javadoc)
*
* @see
- * com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#setSource(org.elasticsearch.action.index
+ * org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#setSource(org.elasticsearch.action.index
* .IndexRequestBuilder, java.lang.Object)
*/
@Override
@@ -62,7 +62,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
/*
* (non-Javadoc)
*
- * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getId(java.lang.Object)
+ * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getId(java.lang.Object)
*/
@Override
protected String getId(T tuple)
@@ -104,7 +104,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
/*
* (non-Javadoc)
*
- * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getIndexName(java.lang.Object)
+ * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getIndexName(java.lang.Object)
*/
@Override
protected String getIndexName(T tuple)
@@ -122,7 +122,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
}
/* (non-Javadoc)
- * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getType(java.lang.Object)
+ * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getType(java.lang.Object)
*/
@Override
protected String getType(T tuple)
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java
index 9f7b66c..614fd59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
import java.io.IOException;
@@ -24,9 +24,9 @@ import javax.validation.constraints.NotNull;
import org.elasticsearch.action.percolate.PercolateResponse;
-import com.datatorrent.common.util.BaseOperator;
import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.api.DefaultOutputPort;
+import com.datatorrent.common.util.BaseOperator;
import com.datatorrent.netlet.util.DTThrowable;
/**
@@ -50,8 +50,8 @@ public class ElasticSearchPercolatorOperator extends BaseOperator
protected transient ElasticSearchPercolatorStore store;
public final transient DefaultOutputPort<PercolateResponse> outputPort = new DefaultOutputPort<PercolateResponse>();
- public final transient DefaultInputPort<Object> inputPort = new DefaultInputPort<Object>() {
-
+ public final transient DefaultInputPort<Object> inputPort = new DefaultInputPort<Object>()
+ {
/*
* (non-Javadoc)
*
@@ -61,7 +61,7 @@ public class ElasticSearchPercolatorOperator extends BaseOperator
public void process(Object tuple)
{
- PercolateResponse response = store.percolate(new String[] { indexName }, documentType, tuple);
+ PercolateResponse response = store.percolate(new String[] {indexName}, documentType, tuple);
outputPort.emit(response);
}
};
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java
index c13c025..19639c7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
import java.io.IOException;
@@ -58,7 +58,8 @@ public class ElasticSearchPercolatorStore extends ElasticSearchConnectable
}
}
- public PercolateResponse percolate(String[] indexNames, String documentType, Object tuple){
+ public PercolateResponse percolate(String[] indexNames, String documentType, Object tuple)
+ {
XContentBuilder docBuilder;
try {
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java
index d89d53f..1c74c32 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java
@@ -17,4 +17,4 @@
* under the License.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java
index c377b96..b7e8820 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java
@@ -16,12 +16,17 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.io.IOException;
import java.util.ArrayList;
import java.util.List;
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
+import org.apache.apex.malhar.lib.db.cache.CacheStore;
+import org.apache.apex.malhar.lib.db.cache.CacheStore.ExpiryType;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.FieldInfo.SupportType;
import org.apache.hadoop.classification.InterfaceStability;
import com.esotericsoftware.kryo.NotNull;
@@ -29,11 +34,6 @@ import com.esotericsoftware.kryo.NotNull;
import com.datatorrent.api.Context;
import com.datatorrent.api.Operator;
import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.db.cache.CacheManager;
-import com.datatorrent.lib.db.cache.CacheStore;
-import com.datatorrent.lib.db.cache.CacheStore.ExpiryType;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.FieldInfo.SupportType;
/**
* Base class for Enrichment Operator. Subclasses should provide implementation to getKey and convert.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java
index 845a756..62d11a0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java
@@ -16,15 +16,14 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.util.List;
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
+import org.apache.apex.malhar.lib.util.FieldInfo;
import org.apache.hadoop.classification.InterfaceStability;
-import com.datatorrent.lib.db.cache.CacheManager;
-import com.datatorrent.lib.util.FieldInfo;
-
/**
* Interface for store to be used in enrichment
*
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java
index 3121cf1..aa4a8bd 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.io.IOException;
import java.util.List;
@@ -30,14 +30,13 @@ import org.supercsv.cellprocessor.ift.CellProcessor;
import org.supercsv.io.CsvMapReader;
import org.supercsv.prefs.CsvPreference;
+import org.apache.apex.malhar.contrib.parser.CellProcessorBuilder;
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema;
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema.Field;
+import org.apache.apex.malhar.lib.util.ReusableStringReader;
import org.apache.commons.lang.StringUtils;
import org.apache.hadoop.classification.InterfaceStability;
-import com.datatorrent.contrib.parser.CellProcessorBuilder;
-import com.datatorrent.contrib.parser.DelimitedSchema;
-import com.datatorrent.contrib.parser.DelimitedSchema.Field;
-import com.datatorrent.lib.util.ReusableStringReader;
-
/**
* This implementation of {@link FSLoader} is used to load data from delimited
* file.User needs to provide a schema as a string specified in a json format as
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java
index e04d6c4..cc23cb2 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.io.BufferedReader;
import java.io.IOException;
@@ -28,6 +28,8 @@ import java.util.Map;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
+import org.apache.apex.malhar.lib.util.FieldInfo;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
@@ -38,8 +40,6 @@ import com.esotericsoftware.kryo.NotNull;
import com.google.common.collect.Maps;
import com.datatorrent.api.Component;
-import com.datatorrent.lib.db.cache.CacheManager;
-import com.datatorrent.lib.util.FieldInfo;
/**
* This implementation of {@link BackendLoader} loads the data from a given file
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java
index 2f54f66..e855c07 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.text.ParseException;
import java.text.SimpleDateFormat;
@@ -30,6 +30,8 @@ import javax.validation.constraints.NotNull;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.contrib.parser.AbstractCsvParser.FIELD_TYPE;
+import org.apache.apex.malhar.contrib.parser.AbstractCsvParser.Field;
import org.apache.commons.lang3.StringUtils;
import org.apache.hadoop.classification.InterfaceStability;
@@ -39,9 +41,6 @@ import com.univocity.parsers.fixed.FixedWidthFields;
import com.univocity.parsers.fixed.FixedWidthParser;
import com.univocity.parsers.fixed.FixedWidthParserSettings;
-import com.datatorrent.contrib.parser.AbstractCsvParser.FIELD_TYPE;
-import com.datatorrent.contrib.parser.AbstractCsvParser.Field;
-
/**
* This implementation of {@link FSLoader} is used to load data from fixed width
* file.User needs to set {@link FixedWidthFSLoader#fieldDescription} to specify
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java
index ce02c08..0ed9e22 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.sql.PreparedStatement;
import java.sql.ResultSet;
@@ -26,12 +26,12 @@ import java.util.ArrayList;
import java.util.List;
import java.util.Map;
+import org.apache.apex.malhar.lib.db.jdbc.JdbcStore;
+import org.apache.apex.malhar.lib.util.FieldInfo;
import org.apache.commons.collections.CollectionUtils;
import org.apache.hadoop.classification.InterfaceStability;
import com.google.common.collect.Lists;
-import com.datatorrent.lib.db.jdbc.JdbcStore;
-import com.datatorrent.lib.util.FieldInfo;
/**
* <p>HBaseLoader extends from {@link JdbcStore} uses JDBC to connect and implements BackendLoaders interface.</p> <br/>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java
index 183d07a..1347100 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.io.IOException;
import java.util.Map;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java
index 6e7b5f7..d3d8df4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java
@@ -16,21 +16,21 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
+import org.apache.apex.malhar.lib.util.FieldInfo;
import org.apache.hadoop.classification.InterfaceStability;
import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.lib.util.FieldInfo;
/**
* This class takes a HashMap tuple as input and extracts value of the lookupKey configured
- * for this operator. It perform a lookup using {@link com.datatorrent.lib.db.cache.CacheManager} to
+ * for this operator. It perform a lookup using {@link org.apache.apex.malhar.lib.db.cache.CacheManager} to
* find a matching entry and adds the result to the original tuple.
*
* <p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java
index 6d42fec..74f3f97 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java
@@ -16,12 +16,11 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
import org.apache.hadoop.classification.InterfaceStability;
-import com.datatorrent.lib.db.cache.CacheManager;
-
/**
* Null Values Cache Manager. Using this NULL entries can be specified explicitly.
*
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java
index f22acdb..1f979ee 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.lang.reflect.Field;
import java.util.ArrayList;
@@ -26,6 +26,9 @@ import java.util.Map;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
import org.apache.commons.lang3.ClassUtils;
import org.apache.hadoop.classification.InterfaceStability;
@@ -35,13 +38,11 @@ import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.api.DefaultOutputPort;
import com.datatorrent.api.annotation.InputPortFieldAnnotation;
import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
/**
* This class takes a POJO as input and extracts the value of the lookupKey configured
- * for this operator. It perform a lookup using {@link com.datatorrent.lib.db.cache.CacheManager} to
+ * for this operator. It perform a lookup using {@link org.apache.apex.malhar.lib.db.cache.CacheManager} to
* find a matching entry and adds the result to the original tuple.
*
* <p>
@@ -212,7 +213,7 @@ public class POJOEnricher extends AbstractEnricher<Object, Object>
@SuppressWarnings({ "unchecked", "rawtypes" })
private PojoUtils.Setter generateSettersForField(Class<?> klass, String outputFieldName)
- throws NoSuchFieldException, SecurityException
+ throws NoSuchFieldException, SecurityException
{
Field f = klass.getDeclaredField(outputFieldName);
Class c = ClassUtils.primitiveToWrapper(f.getType());
@@ -221,7 +222,7 @@ public class POJOEnricher extends AbstractEnricher<Object, Object>
@SuppressWarnings({ "unchecked", "rawtypes" })
private PojoUtils.Getter generateGettersForField(Class<?> klass, String inputFieldName)
- throws NoSuchFieldException, SecurityException
+ throws NoSuchFieldException, SecurityException
{
Field f = klass.getDeclaredField(inputFieldName);
Class c = ClassUtils.primitiveToWrapper(f.getType());
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java
index c6afbe1..0ae3ee7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java
@@ -16,15 +16,14 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
import java.util.List;
import java.util.Map;
+import org.apache.apex.malhar.lib.util.FieldInfo;
import org.apache.hadoop.classification.InterfaceStability;
-import com.datatorrent.lib.util.FieldInfo;
-
/**
* ReadOnly abstract implementation of BackendLoader.
*
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java
index 2bd0e67..3469039 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.formatter;
+package org.apache.apex.malhar.contrib.formatter;
import java.io.IOException;
import java.io.StringWriter;
@@ -34,14 +34,15 @@ import org.supercsv.io.CsvBeanWriter;
import org.supercsv.io.ICsvBeanWriter;
import org.supercsv.prefs.CsvPreference;
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema;
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema.Field;
+import org.apache.apex.malhar.contrib.parser.Schema.FieldType;
+import org.apache.apex.malhar.lib.formatter.Formatter;
+
import com.google.common.annotations.VisibleForTesting;
import com.datatorrent.api.AutoMetric;
import com.datatorrent.api.Context;
-import com.datatorrent.contrib.parser.DelimitedSchema;
-import com.datatorrent.contrib.parser.DelimitedSchema.Field;
-import com.datatorrent.contrib.parser.Schema.FieldType;
-import com.datatorrent.lib.formatter.Formatter;
import com.datatorrent.netlet.util.DTThrowable;
/**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java
index 497e6e4..d56d471 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java
@@ -16,9 +16,9 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
-import com.datatorrent.lib.db.AbstractKeyValueStoreInputOperator;
+import org.apache.apex.malhar.lib.db.AbstractKeyValueStoreInputOperator;
/**
* This is the base implementation used for geode input adapters. A
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java
index dd0bad2..d3951fc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java
@@ -16,9 +16,9 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
/**
* This is the base implementation of geode output operators. A concrete
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java
index 2152b97..0c01373 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java
@@ -16,9 +16,21 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
+
+import java.io.ByteArrayInputStream;
+import java.io.IOException;
+import java.io.Serializable;
+import java.util.ArrayList;
+import java.util.Iterator;
+import java.util.List;
+import java.util.Map;
+import java.util.Map.Entry;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.StorageAgentKeyValueStore;
-import com.datatorrent.lib.util.StorageAgentKeyValueStore;
import com.esotericsoftware.kryo.Kryo;
import com.esotericsoftware.kryo.io.Input;
import com.esotericsoftware.kryo.io.Output;
@@ -32,17 +44,6 @@ import com.gemstone.gemfire.cache.query.Query;
import com.gemstone.gemfire.cache.query.QueryService;
import com.gemstone.gemfire.cache.query.SelectResults;
import com.google.common.collect.Maps;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import java.io.ByteArrayInputStream;
-import java.io.IOException;
-import java.io.Serializable;
-import java.util.ArrayList;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-import java.util.Map.Entry;
/**
* Geode Store implementation of {@link StorageAgentKeyValueStore} Uses {@link Kryo}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java
index 691c2c1..819d531 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java
@@ -16,14 +16,13 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
import java.io.Serializable;
+import org.apache.apex.malhar.lib.util.AbstractKeyValueStorageAgent;
import org.apache.hadoop.conf.Configuration;
-import com.datatorrent.lib.util.AbstractKeyValueStorageAgent;
-
/**
* Storage Agent implementation which uses {@link GeodeCheckpointStore} for operator
* checkpointing
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java
index c7d22c7..d2001e5 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java
@@ -16,15 +16,14 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.TableInfo;
import org.apache.hadoop.classification.InterfaceStability.Evolving;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
-import com.datatorrent.lib.util.TableInfo;
-
/**
*
* @displayName Geode Output Operator
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java
index d345661..803a83e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
import java.io.IOException;
import java.io.Serializable;
@@ -27,7 +27,7 @@ import java.util.Map;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import org.apache.accumulo.core.client.impl.thrift.ThriftTest.Processor.throwsError;
+import org.apache.apex.malhar.lib.db.KeyValueStore;
import com.gemstone.gemfire.cache.CacheClosedException;
import com.gemstone.gemfire.cache.CacheWriterException;
@@ -43,8 +43,6 @@ import com.gemstone.gemfire.cache.query.QueryInvocationTargetException;
import com.gemstone.gemfire.cache.query.SelectResults;
import com.gemstone.gemfire.cache.query.TypeMismatchException;
-import com.datatorrent.lib.db.KeyValueStore;
-
/**
* Provides the implementation of a Geode store.
* Geode is a distributed in-memory database
@@ -145,7 +143,7 @@ public class GeodeStore implements KeyValueStore, Serializable
if (region == null) {
region = clientCache.getRegion(regionName);
if (region == null) {
- region = clientCache.<Object, Object> createClientRegionFactory(ClientRegionShortcut.PROXY).create(regionName);
+ region = clientCache.<Object, Object>createClientRegionFactory(ClientRegionShortcut.PROXY).create(regionName);
}
}
@@ -164,7 +162,7 @@ public class GeodeStore implements KeyValueStore, Serializable
region = clientCache.getRegion(getRegionName());
if (region == null) {
- region = clientCache.<Object, Object> createClientRegionFactory(ClientRegionShortcut.PROXY).create(
+ region = clientCache.<Object, Object>createClientRegionFactory(ClientRegionShortcut.PROXY).create(
getRegionName());
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java
index 9e948c4..d7aa9d5 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
import java.util.ArrayList;
import java.util.List;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java
index a2c1297..d83768d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
@@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.client.HTable;
import com.datatorrent.netlet.util.DTThrowable;
/**
- * A base implementation of a StoreOutputOperator operator that stores tuples in HBase columns and offers
+ * A base implementation of a StoreOutputOperator operator that stores tuples in HBase columns and offers
* non-transactional append. Subclasses should provide implementation for appending operations. <br>
* <p>
* <br>
@@ -49,16 +49,19 @@ import com.datatorrent.netlet.util.DTThrowable;
* The tuple type
* @since 1.0.2
*/
-public abstract class AbstractHBaseAppendOutputOperator<T> extends AbstractHBaseOutputOperator<T> {
+public abstract class AbstractHBaseAppendOutputOperator<T> extends AbstractHBaseOutputOperator<T>
+{
private static final transient Logger logger = LoggerFactory
.getLogger(AbstractHBaseAppendOutputOperator.class);
- public AbstractHBaseAppendOutputOperator() {
+ public AbstractHBaseAppendOutputOperator()
+ {
store = new HBaseStore();
}
@Override
- public void processTuple(T tuple, HTable table) {
+ public void processTuple(T tuple, HTable table)
+ {
Append append = operationAppend(tuple);
try {
table.append(append);
diff --git a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseOutputOperator.java
similarity index 93%
rename from contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseOutputOperator.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseOutputOperator.java
index b9d3c49..f07f123 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseOutputOperator.java
@@ -16,15 +16,15 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
import org.apache.hadoop.hbase.client.HTable;
import com.datatorrent.api.Operator;
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
/**
- * A base implementation of a HBase output operator that stores tuples in HBase and offers
+ * A base implementation of a HBase output operator that stores tuples in HBase and offers
* non-transactional Insert. Subclasses should provide implementation for specific Hbase operations.
*
* @since 3.8.0
@@ -32,7 +32,7 @@ import com.datatorrent.lib.db.AbstractStoreOutputOperator;
public abstract class AbstractHBaseOutputOperator<T> extends AbstractStoreOutputOperator<T, HBaseStore> implements OutputAdapter.Driver<T>, Operator.CheckpointNotificationListener
{
private transient OutputAdapter<T> outputAdapter;
-
+
public AbstractHBaseOutputOperator()
{
outputAdapter = new OutputAdapter<T>(store, this);
@@ -43,13 +43,13 @@ public abstract class AbstractHBaseOutputOperator<T> extends AbstractStoreOutput
{
outputAdapter.processTuple(tuple);
}
-
+
@Override
public abstract void processTuple(T tuple, HTable table);
/**
* Get the table name for tuple.
- *
+ *
* Implementations can override this method to return the table name where the tuple should be written to.
* Return null to write to default table
* @param tuple The tuple
@@ -84,5 +84,5 @@ public abstract class AbstractHBaseOutputOperator<T> extends AbstractStoreOutput
{
}
-
+
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java
index 2421936..cc5f8e4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.InterruptedIOException;
@@ -48,7 +48,8 @@ import com.datatorrent.netlet.util.DTThrowable;
* The tuple type
* @since 1.0.2
*/
-public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOutputOperator<T> {
+public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOutputOperator<T>
+{
private static final transient Logger logger = LoggerFactory.getLogger(AbstractHBasePutOutputOperator.class);
public AbstractHBasePutOutputOperator()
@@ -56,7 +57,6 @@ public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOut
store = new HBaseStore();
}
- @Override
public void processTuple(T tuple, HTable table)
{
Put put = operationPut(tuple);
@@ -72,5 +72,5 @@ public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOut
}
public abstract Put operationPut(T t);
-
+
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java
index 2fa2146..1747cde 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
@@ -30,7 +30,9 @@ import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.netlet.util.DTThrowable;
/**
- * A base implementation of an AggregateTransactionableStoreOutputOperator operator that stores tuples in HBase columns and provides batch append. Subclasses should provide implementation for appending operations. <br>
+ * A base implementation of an AggregateTransactionableStoreOutputOperator
+ * operator that stores tuples in HBase columns and provides batch append.
+ * Subclasses should provide implementation for appending operations. <br>
* <p>
* <br>
* This class provides a HBase output operator that can be used to store tuples
@@ -45,17 +47,17 @@ import com.datatorrent.netlet.util.DTThrowable;
* guarantee each tuple is written only once to HBase in case the operator is
* restarted from an earlier checkpoint. It only tries to minimize the number of
* duplicates limiting it to the tuples that were processed in the window when
- * the operator shutdown.
- * It supports atleast once and atmost once processing modes.
- * Exactly once is not supported
+ * the operator shutdown. It supports atleast once and atmost once processing
+ * modes. Exactly once is not supported
+ *
* @displayName Abstract HBase Window Append Output
* @category Output
* @tags hbase, append, transactionable, batch
- * @param <T>
- * The tuple type
+ * @param <T> The tuple type
* @since 1.0.2
*/
-public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T> {
+public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T>
+{
private static final transient Logger logger = LoggerFactory.getLogger(AbstractHBaseWindowAppendOutputOperator.class);
private transient ProcessingMode mode;
@@ -74,12 +76,14 @@ public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends Abstrac
this.mode = mode;
}
- public AbstractHBaseWindowAppendOutputOperator() {
+ public AbstractHBaseWindowAppendOutputOperator()
+ {
store = new HBaseWindowStore();
}
@Override
- public void processTuple(T tuple, HTable table) {
+ public void processTuple(T tuple, HTable table)
+ {
try {
Append append = operationAppend(tuple);
table.append(append);
@@ -103,8 +107,8 @@ public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends Abstrac
@Override
public void setup(OperatorContext context)
{
- mode=context.getValue(OperatorContext.PROCESSING_MODE);
- if(mode==ProcessingMode.EXACTLY_ONCE){
+ mode = context.getValue(OperatorContext.PROCESSING_MODE);
+ if (mode == ProcessingMode.EXACTLY_ONCE) {
throw new RuntimeException("This operator only supports atmost once and atleast once processing modes");
}
super.setup(context);
diff --git a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseWindowOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowOutputOperator.java
similarity index 95%
rename from contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseWindowOutputOperator.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowOutputOperator.java
index 665b463..fb21bf9 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseWindowOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowOutputOperator.java
@@ -16,15 +16,15 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.AbstractPassThruTransactionableStoreOutputOperator;
import org.apache.hadoop.hbase.client.HTable;
import com.datatorrent.api.Operator;
-import com.datatorrent.lib.db.AbstractPassThruTransactionableStoreOutputOperator;
/**
* A base implementation of an AggregateTransactionableStoreOutputOperator that stores tuples in HBase
@@ -38,7 +38,7 @@ public abstract class AbstractHBaseWindowOutputOperator<T> extends AbstractPassT
* Both at checkpoint window and end window, flush the tuples as application window may not align with
* end window especially when it is more than one streaming window
*/
-
+
private static final Logger logger = LoggerFactory.getLogger(AbstractHBaseWindowOutputOperator.class);
private transient OutputAdapter<T> outputAdapter;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java
index 70fbeeb..e6c65fc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
@@ -30,7 +30,9 @@ import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.netlet.util.DTThrowable;
/**
- * A base implementation of an AggregateTransactionableStoreOutputOperator operator for storing tuples in HBase rows and provides a batch put.Subclasses should provide implementation for put operation. <br>
+ * A base implementation of an AggregateTransactionableStoreOutputOperator
+ * operator for storing tuples in HBase rows and provides a batch put.Subclasses
+ * should provide implementation for put operation. <br>
* <p>
* <br>
* This class provides a HBase output operator that can be used to store tuples
@@ -45,20 +47,20 @@ import com.datatorrent.netlet.util.DTThrowable;
* guarantee each tuple is written only once to HBase in case the operator is
* restarted from an earlier checkpoint. It only tries to minimize the number of
* duplicates limiting it to the tuples that were processed in the window when
- * the operator shutdown.
- * It supports atleast once and atmost once processing modes.
- * Exactly once is not supported
+ * the operator shutdown. It supports atleast once and atmost once processing
+ * modes. Exactly once is not supported
+ *
* @displayName Abstract HBase Window Put Output
* @category Output
* @tags hbase, put, transactionable, batch
- * @param <T>
- * The tuple type
+ * @param <T> The tuple type
* @since 1.0.2
*/
-public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T> {
+public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T>
+{
private static final transient Logger logger = LoggerFactory.getLogger(AbstractHBaseWindowPutOutputOperator.class);
private transient ProcessingMode mode;
-
+
@Deprecated
public ProcessingMode getMode()
{
@@ -71,12 +73,14 @@ public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHB
this.mode = mode;
}
- public AbstractHBaseWindowPutOutputOperator() {
+ public AbstractHBaseWindowPutOutputOperator()
+ {
store = new HBaseWindowStore();
}
@Override
- public void processTuple(T tuple, HTable table) {
+ public void processTuple(T tuple, HTable table)
+ {
try {
Put put = operationPut(tuple);
table.put(put);
@@ -86,14 +90,13 @@ public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHB
}
}
-
public abstract Put operationPut(T t) throws IOException;
@Override
public void setup(OperatorContext context)
{
- mode=context.getValue(context.PROCESSING_MODE);
- if(mode==ProcessingMode.EXACTLY_ONCE){
+ mode = context.getValue(context.PROCESSING_MODE);
+ if (mode == ProcessingMode.EXACTLY_ONCE) {
throw new RuntimeException("This operator only supports atmost once and atleast once processing modes");
}
super.setup(context);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java
index 7420569..468dd8c 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
import java.io.StringReader;
@@ -25,27 +25,31 @@ import java.util.HashMap;
import java.util.Map;
import java.util.Map.Entry;
-import org.apache.hadoop.hbase.client.Put;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.supercsv.io.CsvListReader;
import org.supercsv.io.ICsvListReader;
import org.supercsv.prefs.CsvPreference;
+import org.apache.apex.malhar.lib.util.ReusableStringReader;
+import org.apache.hadoop.hbase.client.Put;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.util.ReusableStringReader;
/**
- * Takes a configuration string which tells us about the position of the row, or column. The incoming tuples are inserted accordingly.
+ * Takes a configuration string which tells us about the position of the row, or
+ * column. The incoming tuples are inserted accordingly.
* <p>
+ *
* @displayName HBase Csv Mapping Put
* @category Output
* @tags hbase, csv, put, String
* @since 1.0.4
*/
-public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOperator<String> {
- private class ColDef {
+public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOperator<String>
+{
+ private class ColDef
+ {
String colFam;
String colName;
}
@@ -58,23 +62,26 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
private transient ArrayList<String> csvLineList = new ArrayList<String>();
private String mappingString;
- public void setMappingString(String mappingString) {
+ public void setMappingString(String mappingString)
+ {
this.mappingString = mappingString;
}
@Override
- public Put operationPut(String t) throws IOException {
+ public Put operationPut(String t) throws IOException
+ {
return parseLine(t);
}
- public void parseMapping() {
+ public void parseMapping()
+ {
ICsvListReader listReader = null;
StringReader sr = null;
ArrayList<String> csvList = new ArrayList<String>();
try {
sr = new StringReader(mappingString);
listReader = new CsvListReader(sr,CsvPreference.STANDARD_PREFERENCE);
- csvList = (ArrayList<String>) listReader.read();
+ csvList = (ArrayList<String>)listReader.read();
} catch (IOException e) {
logger.error("Cannot read the mapping string", e);
DTThrowable.rethrow(e);
@@ -89,9 +96,9 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
}
for (int index = 0; index < csvList.size(); index++) {
String value = csvList.get(index);
- if (value.equals("row"))
+ if (value.equals("row")) {
rowIndex = index;
- else {
+ } else {
ColDef c = new ColDef();
c.colFam = value.substring(0, value.indexOf('.'));
c.colName = value.substring(value.indexOf('.') + 1);
@@ -100,11 +107,12 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
}
}
- public Put parseLine(String s) {
+ public Put parseLine(String s)
+ {
Put put = null;
try {
lineSr.open(s);
- csvLineList = (ArrayList<String>) lineListReader.read();
+ csvLineList = (ArrayList<String>)lineListReader.read();
} catch (IOException e) {
logger.error("Cannot read the property string", e);
DTThrowable.rethrow(e);
@@ -120,15 +128,17 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
}
@Override
- public void setup(OperatorContext context) {
+ public void setup(OperatorContext context)
+ {
super.setup(context);
parseMapping();
lineListReader = new CsvListReader(lineSr,
- CsvPreference.STANDARD_PREFERENCE);
+ CsvPreference.STANDARD_PREFERENCE);
}
@Override
- public void teardown() {
+ public void teardown()
+ {
super.teardown();
try {
lineSr.close();
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java
index 6a34a91..23fa2c5 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java
@@ -16,96 +16,83 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
+import org.apache.apex.malhar.lib.util.FieldInfo;
import org.apache.hadoop.hbase.util.Bytes;
-import com.datatorrent.lib.util.FieldInfo;
-
/**
* @since 3.1.0
*/
public class HBaseFieldInfo extends FieldInfo
{
- private String familyName;
-
- public HBaseFieldInfo()
- {
- }
-
- public HBaseFieldInfo( String columnName, String columnExpression, SupportType type, String familyName )
- {
- super( columnName, columnExpression, type );
- setFamilyName( familyName );
- }
-
- public String getFamilyName()
- {
- return familyName;
- }
-
- public void setFamilyName(String familyName)
- {
- this.familyName = familyName;
- }
+ private String familyName;
- public byte[] toBytes( Object value )
- {
- final SupportType type = getType();
- switch( type )
- {
- case BOOLEAN:
- return Bytes.toBytes( (Boolean)value );
-
- case SHORT:
- return Bytes.toBytes( (Short)value );
-
- case INTEGER:
- return Bytes.toBytes( (Integer)value );
-
- case LONG:
- return Bytes.toBytes( (Long)value );
+ public HBaseFieldInfo()
+ {
+ }
- case FLOAT:
- return Bytes.toBytes( (Float)value );
+ public HBaseFieldInfo( String columnName, String columnExpression, SupportType type, String familyName )
+ {
+ super( columnName, columnExpression, type );
+ setFamilyName( familyName );
+ }
- case DOUBLE:
- return Bytes.toBytes( (Double)value );
+ public String getFamilyName()
+ {
+ return familyName;
+ }
- case STRING:
- return Bytes.toBytes( (String)value );
- }
- throw new IllegalArgumentException( "Unsupported type: " + type );
- }
+ public void setFamilyName(String familyName)
+ {
+ this.familyName = familyName;
+ }
- public Object toValue( byte[] bytes )
- {
+ public byte[] toBytes( Object value )
+ {
final SupportType type = getType();
- switch( type )
- {
- case BOOLEAN:
- return Bytes.toBoolean( bytes );
-
- case SHORT:
- return Bytes.toShort( bytes );
-
- case INTEGER:
- return Bytes.toInt( bytes );
-
- case LONG:
- return Bytes.toLong( bytes );
-
- case FLOAT:
- return Bytes.toFloat( bytes );
-
- case DOUBLE:
- return Bytes.toDouble( bytes );
+ switch (type) {
+ case BOOLEAN:
+ return Bytes.toBytes( (Boolean)value );
+ case SHORT:
+ return Bytes.toBytes( (Short)value );
+ case INTEGER:
+ return Bytes.toBytes( (Integer)value );
+ case LONG:
+ return Bytes.toBytes( (Long)value );
+ case FLOAT:
+ return Bytes.toBytes( (Float)value );
+ case DOUBLE:
+ return Bytes.toBytes( (Double)value );
+ case STRING:
+ return Bytes.toBytes( (String)value );
+ default:
+ throw new IllegalArgumentException("Unsupported type: " + type);
+ }
+ }
- case STRING:
- return Bytes.toString( bytes );
+ public Object toValue( byte[] bytes )
+ {
+ final SupportType type = getType();
+ switch (type) {
+ case BOOLEAN:
+ return Bytes.toBoolean( bytes );
+ case SHORT:
+ return Bytes.toShort( bytes );
+ case INTEGER:
+ return Bytes.toInt( bytes );
+ case LONG:
+ return Bytes.toLong( bytes );
+ case FLOAT:
+ return Bytes.toFloat( bytes );
+ case DOUBLE:
+ return Bytes.toDouble( bytes );
+ case STRING:
+ return Bytes.toString( bytes );
+ default:
+ throw new IllegalArgumentException("Unsupported type: " + type);
}
- throw new IllegalArgumentException( "Unsupported type: " + type );
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java
index fea56a0..a49399e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java
@@ -16,12 +16,12 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.util.List;
-import com.datatorrent.lib.util.FieldValueGenerator;
-import com.datatorrent.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.FieldValueGenerator;
+import org.apache.apex.malhar.lib.util.PojoUtils;
/**
* A {@link FieldValueGenerator} implementation for {@link HBaseFieldInfo}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java
index 37270d5..12608c8 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Get;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java
index 6f11621..82e50b4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java
@@ -16,10 +16,11 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
import com.datatorrent.api.InputOperator;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
/**
* A base implementation of hbase input operator which derives from HBaseOperatorBase. <br>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java
index d35281c..5fd97e1 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java
@@ -16,45 +16,48 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
+
+import java.io.IOException;
+import java.io.StringReader;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.Map;
+import javax.validation.constraints.NotNull;
-import com.datatorrent.api.Context.OperatorContext;
-import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.util.ReusableStringReader;
-import org.apache.hadoop.hbase.client.Put;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.supercsv.io.CsvListReader;
import org.supercsv.io.ICsvListReader;
import org.supercsv.prefs.CsvPreference;
+import org.apache.apex.malhar.lib.util.ReusableStringReader;
+import org.apache.hadoop.hbase.client.Put;
-import javax.validation.constraints.NotNull;
-import java.io.IOException;
-import java.io.StringReader;
-import java.util.ArrayList;
-import java.util.HashMap;
-import java.util.Map;
+import com.datatorrent.api.Context.OperatorContext;
+import com.datatorrent.netlet.util.DTThrowable;
/**
- * An implementation of HBase window put operator that inserts a string of key value pairs which are mapped to corresponding rows, columns.
+ * An implementation of HBase window put operator that inserts a string of key
+ * value pairs which are mapped to corresponding rows, columns.
* <p>
* Accepts a string of key value pairs containing the data to be inserted.These
* are mapped to corresponding rows,column families and columns using a property
* file and inserted into hbase Example: input string will be of the form
- * name="xyz", st="patrickhenry" ,ct="fremont", sa="california", the properties will contain
- * properties of form name=row, sa=address.street, ct=address.city, sa=address.state.
- * With the above mapping a row xyz is created. The value patrickhenry is inserted into
- * columnfamily address and column street of row xyz. Other values are inserted
- * similarly.
+ * name="xyz", st="patrickhenry" ,ct="fremont", sa="california", the properties
+ * will contain properties of form name=row, sa=address.street, ct=address.city,
+ * sa=address.state. With the above mapping a row xyz is created. The value
+ * patrickhenry is inserted into columnfamily address and column street of row
+ * xyz. Other values are inserted similarly.
*
* @displayName HBase Name Value Csv Put
* @category Output
* @tags csv, hbase, put
* @since 1.0.2
*/
-public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOperator<String> {
-
- private class ColDef {
+public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOperator<String>
+{
+ private class ColDef
+ {
String colFam;
String colName;
}
@@ -66,26 +69,29 @@ public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOp
private transient Map<String, ColDef> colMap = new HashMap<String, ColDef>();
private transient Map<String, String> linemap = new HashMap<String, String>();
private transient ICsvListReader lineListReader = null;
- private transient ReusableStringReader lineSr=new ReusableStringReader();
+ private transient ReusableStringReader lineSr = new ReusableStringReader();
private transient ArrayList<String> csvLineList = new ArrayList<String>();
- public void setMapping(String mapping) {
+ public void setMapping(String mapping)
+ {
this.mapping = mapping;
}
@Override
- public Put operationPut(String t) {
+ public Put operationPut(String t)
+ {
return parseLine(t);
}
- public void parseMapping() {
+ public void parseMapping()
+ {
ICsvListReader listReader = null;
- StringReader sr=null;
+ StringReader sr = null;
ArrayList<String> csvList = new ArrayList<String>();
try {
- sr=new StringReader(mapping);
+ sr = new StringReader(mapping);
listReader = new CsvListReader(sr,CsvPreference.STANDARD_PREFERENCE);
- csvList = (ArrayList<String>) listReader.read();
+ csvList = (ArrayList<String>)listReader.read();
} catch (IOException e) {
logger.error("Cannot read the mapping string", e);
DTThrowable.rethrow(e);
@@ -110,11 +116,12 @@ public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOp
}
}
- public Put parseLine(String s) {
+ public Put parseLine(String s)
+ {
Put put = null;
try {
lineSr.open(s);
- csvLineList = (ArrayList<String>) lineListReader.read();
+ csvLineList = (ArrayList<String>)lineListReader.read();
} catch (IOException e) {
logger.error("Cannot read the property string", e);
DTThrowable.rethrow(e);
@@ -144,14 +151,16 @@ public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOp
}
@Override
- public void setup(OperatorContext context) {
+ public void setup(OperatorContext context)
+ {
super.setup(context);
parseMapping();
lineListReader = new CsvListReader(lineSr,CsvPreference.STANDARD_PREFERENCE);
}
@Override
- public void teardown() {
+ public void teardown()
+ {
super.teardown();
try {
lineSr.close();
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java
index 8bed824..04a51b4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
import org.apache.hadoop.conf.Configuration;
@@ -103,7 +103,8 @@ public class HBaseOperatorBase
* Get the configuration.
* @return The configuration
*/
- public Configuration getConfiguration() {
+ public Configuration getConfiguration()
+ {
return configuration;
}
@@ -113,7 +114,8 @@ public class HBaseOperatorBase
* should be called before using the configuration or table.
* @throws IOException
*/
- protected void setupConfiguration() throws IOException {
+ protected void setupConfiguration() throws IOException
+ {
configuration = HBaseConfiguration.create();
configuration.set("hbase.zookeeper.quorum", zookeeperQuorum);
configuration.set("hbase.zookeeper.property.clientPort", "" + zookeeperClientPort);
@@ -127,7 +129,8 @@ public class HBaseOperatorBase
* @return The HBase table if configuration setup was successful, null otherwise
* @throws IOException
*/
- protected HTable getTable() throws IOException {
+ protected HTable getTable() throws IOException
+ {
return table;
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java
index e459ec7..cb36dbf 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java
@@ -16,12 +16,18 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.util.List;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.util.FieldValueGenerator;
+import org.apache.apex.malhar.lib.util.FieldValueGenerator.ValueConverter;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Setter;
+import org.apache.apex.malhar.lib.util.TableInfo;
import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.CellUtil;
@@ -29,14 +35,8 @@ import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.util.Bytes;
-import com.datatorrent.lib.util.FieldValueGenerator;
-import com.datatorrent.lib.util.FieldValueGenerator.ValueConverter;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Setter;
-import com.datatorrent.lib.util.TableInfo;
import com.datatorrent.api.Context;
import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.api.Context.OperatorContext;
/**
* HBasePOJOInputOperator reads data from a HBase store, converts it to a POJO and puts it on the output port.
@@ -96,20 +96,20 @@ public class HBasePOJOInputOperator extends HBaseScanOperator<Object>
{
try {
String readRow = Bytes.toString(result.getRow());
- if( readRow.equals( getLastReadRow() )) {
+ if ( readRow.equals( getLastReadRow() )) {
return null;
}
Object instance = pojoType.newInstance();
rowSetter.set(instance, readRow);
- List<Cell> cells = result.listCells();
- for (Cell cell : cells) {
- String columnName = Bytes.toString(CellUtil.cloneQualifier(cell));
- String columnFamily = Bytes.toString(CellUtil.cloneFamily(cell));
+ List<Cell> cells = result.listCells();
+ for (Cell cell : cells) {
+ String columnName = Bytes.toString(CellUtil.cloneQualifier(cell));
+ String columnFamily = Bytes.toString(CellUtil.cloneFamily(cell));
byte[] value = CellUtil.cloneValue(cell);
- ((HBaseFieldValueGenerator)fieldValueGenerator).setColumnValue(instance, columnName, columnFamily, value,
- valueConverter);
+ ((HBaseFieldValueGenerator)fieldValueGenerator).setColumnValue(instance, columnName, columnFamily, value,
+ valueConverter);
}
setLastReadRow(readRow);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java
index 37c962b..d70f3c7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java
@@ -16,20 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.util.List;
+import org.apache.apex.malhar.lib.util.FieldValueGenerator;
+import org.apache.apex.malhar.lib.util.FieldValueGenerator.FieldValueHandler;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.TableInfo;
import org.apache.hadoop.classification.InterfaceStability.Evolving;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.util.Bytes;
-import com.datatorrent.lib.util.FieldValueGenerator;
-import com.datatorrent.lib.util.FieldValueGenerator.FieldValueHandler;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
-import com.datatorrent.lib.util.TableInfo;
-
/**
* @displayName HBase Output Operator
* @category Output
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java
index b694e67..f525058 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java
@@ -16,36 +16,38 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
import java.util.Queue;
-import org.apache.hadoop.hbase.KeyValue;
-import org.apache.hadoop.hbase.client.Scan;
+
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.ResultScanner;
+import org.apache.hadoop.hbase.client.Scan;
+import com.google.common.collect.Queues;
import com.datatorrent.api.AutoMetric;
import com.datatorrent.api.Context;
-import com.datatorrent.api.Operator;
import com.datatorrent.api.Context.OperatorContext;
-import com.google.common.collect.Queues;
-
-import org.apache.hadoop.hbase.client.HTable;
-import org.apache.hadoop.hbase.client.Result;
-import org.apache.hadoop.hbase.client.ResultScanner;
+import com.datatorrent.api.Operator;
/**
- * A base implementation of hbase input operator that retrieves tuples from HBase columns and provides scan operation. <br>
+ * A base implementation of hbase input operator that retrieves tuples from
+ * HBase columns and provides scan operation. <br>
* <p>
* <br>
- * This class provides a HBase input operator that can be used to retrieve tuples from rows in a
- * HBase table. The class should be extended by the end-operator developer. The extending class should
- * implement operationScan and getTuple methods. The operationScan method should provide a HBase Scan
- * metric object that specifies where to retrieve the tuple information from the table. The getTuple method
- * should map the contents of a Result from the Scan result to a tuple.<br>
+ * This class provides a HBase input operator that can be used to retrieve
+ * tuples from rows in a HBase table. The class should be extended by the
+ * end-operator developer. The extending class should implement operationScan
+ * and getTuple methods. The operationScan method should provide a HBase Scan
+ * metric object that specifies where to retrieve the tuple information from the
+ * table. The getTuple method should map the contents of a Result from the Scan
+ * result to a tuple.<br>
*
* <br>
+ *
* @displayName HBase Scan
* @category Output
* @tags hbase, scan, input operator
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java
index 2b0ffa7..d46bc6f 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
import java.io.InterruptedIOException;
@@ -27,6 +27,7 @@ import javax.validation.constraints.Min;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.Connectable;
import org.apache.commons.lang.ArrayUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.HBaseConfiguration;
@@ -41,7 +42,6 @@ import com.google.common.cache.LoadingCache;
import com.google.common.cache.RemovalListener;
import com.google.common.cache.RemovalNotification;
-import com.datatorrent.lib.db.Connectable;
/**
* A {@link Connectable} that uses HBase to connect to stores and implements Connectable interface.
* <p>
@@ -50,15 +50,15 @@ import com.datatorrent.lib.db.Connectable;
* @tags store, hbase
* @since 1.0.2
*/
-public class HBaseStore implements Connectable {
-
+public class HBaseStore implements Connectable
+{
public static final String USER_NAME_SPECIFIER = "%USER_NAME%";
private static final Logger logger = LoggerFactory.getLogger(HBaseStore.class);
private String zookeeperQuorum;
private int zookeeperClientPort;
-
+
// Default table name if specified
protected String tableName;
@@ -67,7 +67,7 @@ public class HBaseStore implements Connectable {
// Default interval 30 min
protected long reloginCheckInterval = 30 * 60 * 1000;
protected transient Thread loginRenewer;
- private volatile transient boolean doRelogin;
+ private transient volatile boolean doRelogin;
protected transient HTable table;
// Multi - table
@@ -76,13 +76,14 @@ public class HBaseStore implements Connectable {
@Min(1)
protected int maxOpenTables = Integer.MAX_VALUE;
-
+
/**
* Get the zookeeper quorum location.
*
* @return The zookeeper quorum location
*/
- public String getZookeeperQuorum() {
+ public String getZookeeperQuorum()
+ {
return zookeeperQuorum;
}
@@ -92,7 +93,8 @@ public class HBaseStore implements Connectable {
* @param zookeeperQuorum
* The zookeeper quorum location
*/
- public void setZookeeperQuorum(String zookeeperQuorum) {
+ public void setZookeeperQuorum(String zookeeperQuorum)
+ {
this.zookeeperQuorum = zookeeperQuorum;
}
@@ -101,7 +103,8 @@ public class HBaseStore implements Connectable {
*
* @return The zookeeper client port
*/
- public int getZookeeperClientPort() {
+ public int getZookeeperClientPort()
+ {
return zookeeperClientPort;
}
@@ -111,7 +114,8 @@ public class HBaseStore implements Connectable {
* @param zookeeperClientPort
* The zookeeper client port
*/
- public void setZookeeperClientPort(int zookeeperClientPort) {
+ public void setZookeeperClientPort(int zookeeperClientPort)
+ {
this.zookeeperClientPort = zookeeperClientPort;
}
@@ -120,7 +124,8 @@ public class HBaseStore implements Connectable {
*
* @return The HBase table name
*/
- public String getTableName() {
+ public String getTableName()
+ {
return tableName;
}
@@ -130,7 +135,8 @@ public class HBaseStore implements Connectable {
* @param tableName
* The HBase table name
*/
- public void setTableName(String tableName) {
+ public void setTableName(String tableName)
+ {
this.tableName = tableName;
}
@@ -199,8 +205,8 @@ public class HBaseStore implements Connectable {
/**
* Gets the allowedtableNames
- *
- * @return allowedTableNames
+ *
+ * @return allowedTableNames
*/
public String[] getAllowedTableNames()
{
@@ -211,14 +217,15 @@ public class HBaseStore implements Connectable {
{
this.allowedTableNames = allowedTableNames;
}
-
+
/**
* Get the HBase table .
*
* @return The HBase table
* @omitFromUI
*/
- public HTable getTable() {
+ public HTable getTable()
+ {
return table;
}
@@ -230,16 +237,18 @@ public class HBaseStore implements Connectable {
* @return The HBase table
* @omitFromUI
*/
- public HTable getTable(String tableName) {
- if ((tableName == null) || tableName.equals(this.tableName))
+ public HTable getTable(String tableName)
+ {
+ if ((tableName == null) || tableName.equals(this.tableName)) {
return getTable();
+ }
try {
return tableCache.get(tableName);
} catch (Exception e) {
throw Throwables.propagate(e);
}
}
-
+
public void flushTables() throws InterruptedIOException, RetriesExhaustedWithDetailsException
{
if (table != null) {
@@ -249,7 +258,7 @@ public class HBaseStore implements Connectable {
flushTable(entry.getValue());
}
}
-
+
protected void flushTable(HTable table) throws InterruptedIOException, RetriesExhaustedWithDetailsException
{
table.flushCommits();
@@ -260,7 +269,8 @@ public class HBaseStore implements Connectable {
*
* @return The configuration
*/
- public Configuration getConfiguration() {
+ public Configuration getConfiguration()
+ {
return configuration;
}
@@ -270,14 +280,16 @@ public class HBaseStore implements Connectable {
* @param configuration
* The configuration
*/
- public void setConfiguration(Configuration configuration) {
+ public void setConfiguration(Configuration configuration)
+ {
this.configuration = configuration;
}
protected transient Configuration configuration;
@Override
- public void connect() throws IOException {
+ public void connect() throws IOException
+ {
if ((principal != null) && (keytabPath != null)) {
String lprincipal = evaluateProperty(principal);
String lkeytabPath = evaluateProperty(keytabPath);
@@ -317,12 +329,12 @@ public class HBaseStore implements Connectable {
if (zookeeperClientPort != 0) {
configuration.set("hbase.zookeeper.property.clientPort", "" + zookeeperClientPort);
}
-
+
// Connect to default table if specified
if (tableName != null) {
table = connectTable(tableName);
}
-
+
CacheLoader<String, HTable> cacheLoader = new CacheLoader<String, HTable>()
{
@Override
@@ -331,7 +343,7 @@ public class HBaseStore implements Connectable {
return loadTable(key);
}
};
-
+
RemovalListener<String, HTable> removalListener = new RemovalListener<String, HTable>()
{
@Override
@@ -340,9 +352,9 @@ public class HBaseStore implements Connectable {
unloadTable(notification.getValue());
}
};
-
+
int maxCacheSize = (tableName == null) ? maxOpenTables : (maxOpenTables - 1);
-
+
tableCache = CacheBuilder.<String, HTable>newBuilder().maximumSize(maxCacheSize).removalListener(removalListener).build(cacheLoader);
}
@@ -362,7 +374,7 @@ public class HBaseStore implements Connectable {
logger.warn("Could not close table", e);
}
}
-
+
protected HTable connectTable(String tableName) throws IOException
{
HTable table = new HTable(configuration, tableName);
@@ -373,13 +385,14 @@ public class HBaseStore implements Connectable {
private String evaluateProperty(String property) throws IOException
{
if (property.contains(USER_NAME_SPECIFIER)) {
- property = property.replaceAll(USER_NAME_SPECIFIER, UserGroupInformation.getLoginUser().getShortUserName());
+ property = property.replaceAll(USER_NAME_SPECIFIER, UserGroupInformation.getLoginUser().getShortUserName());
}
return property;
}
@Override
- public void disconnect() throws IOException {
+ public void disconnect() throws IOException
+ {
if (loginRenewer != null) {
doRelogin = false;
loginRenewer.interrupt();
@@ -392,7 +405,8 @@ public class HBaseStore implements Connectable {
}
@Override
- public boolean isConnected() {
+ public boolean isConnected()
+ {
// not applicable to hbase
return false;
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java
index 2c9f539..38d6195 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.IOException;
import java.io.InterruptedIOException;
@@ -24,6 +24,7 @@ import java.io.InterruptedIOException;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.KeyValue;
@@ -36,8 +37,8 @@ import org.apache.hadoop.hbase.util.Bytes;
import com.google.common.base.Throwables;
-import com.datatorrent.lib.db.TransactionableStore;
import com.datatorrent.netlet.util.DTThrowable;
+
/**
* <p>Provides transaction support to the operators by implementing TransactionableStore abstract methods. </p>
* <p>
@@ -49,7 +50,8 @@ import com.datatorrent.netlet.util.DTThrowable;
* @tags store, transactional
* @since 1.0.2
*/
-public class HBaseWindowStore extends HBaseStore implements TransactionableStore {
+public class HBaseWindowStore extends HBaseStore implements TransactionableStore
+{
private static final transient Logger logger = LoggerFactory.getLogger(HBaseWindowStore.class);
private static final String DEFAULT_ROW_NAME = "HBaseOperator_row";
private static final String DEFAULT_COLUMN_FAMILY_NAME = "HBaseOutputOperator_cf";
@@ -63,14 +65,16 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
private transient String lastWindowColumnName;
private transient byte[] lastWindowColumnBytes;
- public HBaseWindowStore() {
+ public HBaseWindowStore()
+ {
rowName = DEFAULT_ROW_NAME;
columnFamilyName = DEFAULT_COLUMN_FAMILY_NAME;
lastWindowColumnName = DEFAULT_LAST_WINDOW_PREFIX_COLUMN_NAME;
constructKeys();
}
- private void constructKeys() {
+ private void constructKeys()
+ {
rowBytes = Bytes.toBytes(rowName);
columnFamilyBytes = Bytes.toBytes(columnFamilyName);
}
@@ -80,7 +84,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
*
* @return The row name
*/
- public String getRowName() {
+ public String getRowName()
+ {
return rowName;
}
@@ -90,7 +95,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
* @param rowName
* The row name
*/
- public void setRowName(String rowName) {
+ public void setRowName(String rowName)
+ {
this.rowName = rowName;
constructKeys();
}
@@ -100,7 +106,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
*
* @return The column family name
*/
- public String getColumnFamilyName() {
+ public String getColumnFamilyName()
+ {
return columnFamilyName;
}
@@ -110,13 +117,15 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
* @param columnFamilyName
* The column family name
*/
- public void setColumnFamilyName(String columnFamilyName) {
+ public void setColumnFamilyName(String columnFamilyName)
+ {
this.columnFamilyName = columnFamilyName;
constructKeys();
}
@Override
- public void connect() throws IOException {
+ public void connect() throws IOException
+ {
super.connect();
HTableDescriptor tdesc = table.getTableDescriptor();
if (!tdesc.hasFamily(columnFamilyBytes)) {
@@ -133,13 +142,14 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
}
@Override
- public void beginTransaction() {
+ public void beginTransaction()
+ {
// HBase does not support transactions so this method left empty
-
}
@Override
- public void commitTransaction() {
+ public void commitTransaction()
+ {
try {
flushTables();
} catch (InterruptedIOException | RetriesExhaustedWithDetailsException e) {
@@ -148,19 +158,21 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
}
@Override
- public void rollbackTransaction() {
+ public void rollbackTransaction()
+ {
// HBase does not support transactions so this method left empty
-
}
@Override
- public boolean isInTransaction() {
+ public boolean isInTransaction()
+ {
// HBase does not support transactions so this method left empty
return false;
}
@Override
- public long getCommittedWindowId(String appId, int operatorId) {
+ public long getCommittedWindowId(String appId, int operatorId)
+ {
byte[] value = null;
try {
String columnKey = appId + "_" + operatorId + "_" + lastWindowColumnName;
@@ -183,12 +195,14 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
if (value != null) {
long longval = Bytes.toLong(value);
return longval;
- } else
+ } else {
return -1;
+ }
}
@Override
- public void storeCommittedWindowId(String appId, int operatorId,long windowId) {
+ public void storeCommittedWindowId(String appId, int operatorId,long windowId)
+ {
byte[] WindowIdBytes = Bytes.toBytes(windowId);
String columnKey = appId + "_" + operatorId + "_" + lastWindowColumnName;
lastWindowColumnBytes = Bytes.toBytes(columnKey);
@@ -206,7 +220,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
}
@Override
- public void removeCommittedWindowId(String appId, int operatorId) {
+ public void removeCommittedWindowId(String appId, int operatorId)
+ {
// Not applicable to hbase
}
diff --git a/contrib/src/main/java/com/datatorrent/contrib/hbase/OutputAdapter.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/OutputAdapter.java
similarity index 98%
rename from contrib/src/main/java/com/datatorrent/contrib/hbase/OutputAdapter.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/OutputAdapter.java
index 31317bd..e503ebb 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/hbase/OutputAdapter.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/OutputAdapter.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
import java.io.InterruptedIOException;
@@ -75,7 +75,9 @@ public class OutputAdapter<T>
interface Driver<T>
{
void processTuple(T tuple, HTable table);
+
String getTableName(T tuple);
+
void errorTuple(T tuple);
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java
index 0ccd6c2..2df5c10 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java
@@ -20,4 +20,4 @@
* HBase operators and utilities.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java
index a6fb42f..997c890 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java
@@ -16,17 +16,23 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.jython;
-
-import com.datatorrent.api.Context.OperatorContext;
-import com.datatorrent.lib.script.ScriptOperator;
-import org.python.core.*;
-import org.python.util.PythonInterpreter;
+package org.apache.apex.malhar.contrib.jython;
import java.util.HashMap;
import java.util.Iterator;
import java.util.Map;
+import org.python.core.PyCode;
+import org.python.core.PyIterator;
+import org.python.core.PyObject;
+import org.python.core.PyStringMap;
+import org.python.core.PyTuple;
+import org.python.util.PythonInterpreter;
+
+import org.apache.apex.malhar.lib.script.ScriptOperator;
+
+import com.datatorrent.api.Context.OperatorContext;
+
/**
* An operator that executes a python script and passes the input as bindings.
* <p></p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java
index 08b3a3e..9fc7de7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java
@@ -17,4 +17,4 @@
* under the License.
*/
@org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.jython;
+package org.apache.apex.malhar.contrib.jython;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java
index b11c8ef..ec082d1 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java
@@ -16,12 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
+import java.nio.ByteBuffer;
+import java.util.HashMap;
+import java.util.Map;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import com.google.common.collect.Sets;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultInputPort;
import com.datatorrent.common.util.Pair;
-import com.google.common.collect.Sets;
import kafka.api.FetchRequest;
import kafka.api.FetchRequestBuilder;
import kafka.javaapi.FetchResponse;
@@ -32,13 +39,6 @@ import kafka.message.Message;
import kafka.message.MessageAndOffset;
import kafka.producer.KeyedMessage;
import kafka.producer.Partitioner;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import java.nio.ByteBuffer;
-import java.util.HashMap;
-import java.util.Map;
-
/**
* This is a base implementation of a Kafka output operator,
@@ -82,7 +82,6 @@ import java.util.Map;
*/
public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends AbstractKafkaOutputOperator<K, V>
{
-
private Map<Integer, Pair<byte[], byte[]>> lastMsgs;
private transient Partitioner partitioner;
@@ -94,9 +93,9 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
{
super.setup(context);
try {
- String className = (String) getConfigProperties().get(KafkaMetadataUtil.PRODUCER_PROP_PARTITIONER);
+ String className = (String)getConfigProperties().get(KafkaMetadataUtil.PRODUCER_PROP_PARTITIONER);
if (className != null) {
- partitioner = (Partitioner) Class.forName(className).newInstance();
+ partitioner = (Partitioner)Class.forName(className).newInstance();
}
} catch (Exception e) {
throw new RuntimeException("Failed to initialize partitioner", e);
@@ -108,22 +107,22 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
/**
* This input port receives tuples that will be written out to Kafka.
*/
- public final transient DefaultInputPort<T> inputPort = new DefaultInputPort<T>() {
+ public final transient DefaultInputPort<T> inputPort = new DefaultInputPort<T>()
+ {
@Override
public void process(T tuple)
{
Pair<K, V> keyValue = tupleToKeyValue(tuple);
int pid = 0;
- if(partitioner!=null){
+ if (partitioner != null) {
pid = partitioner.partition(keyValue.first, partitionNum);
}
Pair<byte[], byte[]> lastMsg = lastMsgs.get(pid);
-
- if(lastMsg == null || compareToLastMsg(keyValue, lastMsg) > 0){
+ if (lastMsg == null || compareToLastMsg(keyValue, lastMsg) > 0) {
getProducer().send(new KeyedMessage<K, V>(getTopic(), keyValue.first, keyValue.second));
- sendCount ++;
+ sendCount++;
} else {
// ignore tuple because kafka has already had the tuple
logger.debug("Ingore tuple " + tuple);
@@ -134,7 +133,6 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
private void initializeLastProcessingOffset()
{
-
// read last received kafka message
TopicMetadata tm = KafkaMetadataUtil.getTopicMetadata(Sets.newHashSet((String)getConfigProperties().get(KafkaMetadataUtil.PRODUCER_PROP_BROKERLIST)), this.getTopic());
@@ -170,9 +168,7 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
key.get(keyBytes);
lastMsgs.put(pm.partitionId(), new Pair<byte[], byte[]>(keyBytes, valueBytes));
}
-
}
-
}
/**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java
index 1218f4a..9130923 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java
@@ -16,7 +16,35 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
+
+import java.io.IOException;
+import java.lang.reflect.Array;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+
+import javax.validation.Valid;
+import javax.validation.constraints.Min;
+import javax.validation.constraints.NotNull;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.KryoCloneUtils;
+import org.apache.apex.malhar.lib.wal.WindowDataManager;
+import org.apache.commons.lang3.StringUtils;
+import org.apache.commons.lang3.tuple.MutablePair;
+import org.apache.commons.lang3.tuple.Pair;
+
+import com.google.common.base.Joiner;
+import com.google.common.collect.Lists;
+import com.google.common.collect.Sets;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.DefaultPartition;
@@ -28,10 +56,6 @@ import com.datatorrent.api.Stats;
import com.datatorrent.api.StatsListener;
import com.datatorrent.api.annotation.OperatorAnnotation;
import com.datatorrent.api.annotation.Stateless;
-import com.datatorrent.lib.util.KryoCloneUtils;
-import com.google.common.base.Joiner;
-import com.google.common.collect.Lists;
-import com.google.common.collect.Sets;
import kafka.api.FetchRequest;
import kafka.api.FetchRequestBuilder;
@@ -42,30 +66,7 @@ import kafka.javaapi.consumer.SimpleConsumer;
import kafka.message.Message;
import kafka.message.MessageAndOffset;
-import org.apache.apex.malhar.lib.wal.WindowDataManager;
-import org.apache.commons.lang3.StringUtils;
-import org.apache.commons.lang3.tuple.MutablePair;
-import org.apache.commons.lang3.tuple.Pair;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import javax.validation.Valid;
-import javax.validation.constraints.Min;
-import javax.validation.constraints.NotNull;
-
-import java.io.IOException;
-import java.lang.reflect.Array;
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.HashMap;
-import java.util.HashSet;
-import java.util.Iterator;
-import java.util.LinkedList;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-
-import static com.datatorrent.contrib.kafka.KafkaConsumer.KafkaMeterStatsUtil.getOffsetsForPartitions;
+import static org.apache.apex.malhar.contrib.kafka.KafkaConsumer.KafkaMeterStatsUtil.getOffsetsForPartitions;
/**
* This is a base implementation of a Kafka input operator, which consumes data from Kafka message bus.
@@ -226,7 +227,8 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
* processing of messages doesn't get stuck.
* @return The maximum for the total size
*/
- public long getMaxTotalMsgSizePerWindow() {
+ public long getMaxTotalMsgSizePerWindow()
+ {
return maxTotalMsgSizePerWindow;
}
@@ -236,7 +238,8 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
*
* @param maxTotalMsgSizePerWindow The maximum for the total size
*/
- public void setMaxTotalMsgSizePerWindow(long maxTotalMsgSizePerWindow) {
+ public void setMaxTotalMsgSizePerWindow(long maxTotalMsgSizePerWindow)
+ {
this.maxTotalMsgSizePerWindow = maxTotalMsgSizePerWindow;
}
@@ -256,7 +259,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
}
this.context = context;
operatorId = context.getId();
- if(consumer instanceof HighlevelKafkaConsumer && !(windowDataManager instanceof WindowDataManager.NoopWindowDataManager)) {
+ if (consumer instanceof HighlevelKafkaConsumer && !(windowDataManager instanceof WindowDataManager.NoopWindowDataManager)) {
throw new RuntimeException("Idempotency is not supported for High Level Kafka Consumer");
}
windowDataManager.setup(context);
@@ -289,7 +292,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
if (recoveredData != null) {
Map<String, List<PartitionMetadata>> pms = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().topic);
if (pms != null) {
- SimpleKafkaConsumer cons = (SimpleKafkaConsumer) getConsumer();
+ SimpleKafkaConsumer cons = (SimpleKafkaConsumer)getConsumer();
// add all partition request in one Fretch request together
FetchRequestBuilder frb = new FetchRequestBuilder().clientId(cons.getClientId());
for (Map.Entry<KafkaPartition, MutablePair<Long, Integer>> rc : recoveredData.entrySet()) {
@@ -299,13 +302,14 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
Iterator<PartitionMetadata> pmIterator = pmsVal.iterator();
PartitionMetadata pm = pmIterator.next();
while (pm.partitionId() != kp.getPartitionId()) {
- if (!pmIterator.hasNext())
+ if (!pmIterator.hasNext()) {
break;
+ }
pm = pmIterator.next();
}
- if (pm.partitionId() != kp.getPartitionId())
+ if (pm.partitionId() != kp.getPartitionId()) {
continue;
-
+ }
Broker bk = pm.leader();
frb.addFetch(consumer.topic, rc.getKey().getPartitionId(), rc.getValue().left, cons.getBufferSize());
@@ -319,13 +323,14 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
emitTuple(kafkaMessage);
offsetStats.put(kp, msg.offset());
count = count + 1;
- if (count.equals(rc.getValue().right))
+ if (count.equals(rc.getValue().right)) {
break;
+ }
}
}
}
}
- if(windowId == windowDataManager.getLargestCompletedWindow()) {
+ if (windowId == windowDataManager.getLargestCompletedWindow()) {
// Start the consumer at the largest recovery window
SimpleKafkaConsumer cons = (SimpleKafkaConsumer)getConsumer();
// Set the offset positions to the consumer
@@ -337,8 +342,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
cons.resetOffset(currentOffsets);
cons.start();
}
- }
- catch (IOException e) {
+ } catch (IOException e) {
throw new RuntimeException("replay", e);
}
}
@@ -354,8 +358,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
if (currentWindowId > windowDataManager.getLargestCompletedWindow()) {
try {
windowDataManager.save(currentWindowRecoveryState, currentWindowId);
- }
- catch (IOException e) {
+ } catch (IOException e) {
throw new RuntimeException("saving recovery", e);
}
}
@@ -397,8 +400,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
try {
windowDataManager.committed(windowId);
- }
- catch (IOException e) {
+ } catch (IOException e) {
throw new RuntimeException("deleting state", e);
}
}
@@ -453,10 +455,10 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
emitTotalMsgSize += message.msg.size();
offsetStats.put(message.kafkaPart, message.offSet);
MutablePair<Long, Integer> offsetAndCount = currentWindowRecoveryState.get(message.kafkaPart);
- if(offsetAndCount == null) {
+ if (offsetAndCount == null) {
currentWindowRecoveryState.put(message.kafkaPart, new MutablePair<Long, Integer>(message.offSet, 1));
} else {
- offsetAndCount.setRight(offsetAndCount.right+1);
+ offsetAndCount.setRight(offsetAndCount.right + 1);
}
}
}
@@ -507,7 +509,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
boolean isInitialParitition = true;
// check if it's the initial partition
- if(partitions.iterator().hasNext()) {
+ if (partitions.iterator().hasNext()) {
isInitialParitition = partitions.iterator().next().getStats() == null;
}
@@ -516,7 +518,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
// initialize the offset
Map<KafkaPartition, Long> initOffset = null;
- if(isInitialParitition && offsetManager !=null){
+ if (isInitialParitition && offsetManager != null) {
initOffset = offsetManager.loadInitialOffsets();
logger.info("Initial offsets: {} ", "{ " + Joiner.on(", ").useForNull("").withKeyValueSeparator(": ").join(initOffset) + " }");
}
@@ -527,95 +529,92 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
switch (strategy) {
- // For the 1 to 1 mapping The framework will create number of operator partitions based on kafka topic partitions
- // Each operator partition will consume from only one kafka partition
- case ONE_TO_ONE:
-
- if (isInitialParitition) {
- lastRepartitionTime = System.currentTimeMillis();
- logger.info("[ONE_TO_ONE]: Initializing partition(s)");
- // get partition metadata for topics.
- // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
- // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
- Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
-
- // initialize the number of operator partitions according to number of kafka partitions
-
- newPartitions = new LinkedList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>();
- for (Map.Entry<String, List<PartitionMetadata>> kp : kafkaPartitions.entrySet()) {
- String clusterId = kp.getKey();
- for (PartitionMetadata pm : kp.getValue()) {
- logger.info("[ONE_TO_ONE]: Create operator partition for cluster {}, topic {}, kafka partition {} ", clusterId, getConsumer().topic, pm.partitionId());
- newPartitions.add(createPartition(Sets.newHashSet(new KafkaPartition(clusterId, consumer.topic, pm.partitionId())), initOffset));
+ // For the 1 to 1 mapping The framework will create number of operator partitions based on kafka topic partitions
+ // Each operator partition will consume from only one kafka partition
+ case ONE_TO_ONE:
+
+ if (isInitialParitition) {
+ lastRepartitionTime = System.currentTimeMillis();
+ logger.info("[ONE_TO_ONE]: Initializing partition(s)");
+ // get partition metadata for topics.
+ // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
+ // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
+ Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
+
+ // initialize the number of operator partitions according to number of kafka partitions
+
+ newPartitions = new LinkedList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>();
+ for (Map.Entry<String, List<PartitionMetadata>> kp : kafkaPartitions.entrySet()) {
+ String clusterId = kp.getKey();
+ for (PartitionMetadata pm : kp.getValue()) {
+ logger.info("[ONE_TO_ONE]: Create operator partition for cluster {}, topic {}, kafka partition {} ", clusterId, getConsumer().topic, pm.partitionId());
+ newPartitions.add(createPartition(Sets.newHashSet(new KafkaPartition(clusterId, consumer.topic, pm.partitionId())), initOffset));
+ }
+ }
+ resultPartitions = newPartitions;
+ numPartitionsChanged = true;
+ } else if (newWaitingPartition.size() != 0) {
+ // add partition for new kafka partition
+ for (KafkaPartition newPartition : newWaitingPartition) {
+ logger.info("[ONE_TO_ONE]: Add operator partition for cluster {}, topic {}, partition {}", newPartition.getClusterId(), getConsumer().topic, newPartition.getPartitionId());
+ partitions.add(createPartition(Sets.newHashSet(newPartition), null));
}
+ newWaitingPartition.clear();
+ resultPartitions = partitions;
+ numPartitionsChanged = true;
}
- resultPartitions = newPartitions;
- numPartitionsChanged = true;
- }
- else if (newWaitingPartition.size() != 0) {
- // add partition for new kafka partition
- for (KafkaPartition newPartition : newWaitingPartition) {
- logger.info("[ONE_TO_ONE]: Add operator partition for cluster {}, topic {}, partition {}", newPartition.getClusterId(), getConsumer().topic, newPartition.getPartitionId());
- partitions.add(createPartition(Sets.newHashSet(newPartition), null));
+ break;
+ // For the 1 to N mapping The initial partition number is defined by stream application
+ // Afterwards, the framework will dynamically adjust the partition and allocate consumers to as less operator partitions as it can
+ // and guarantee the total intake rate for each operator partition is below some threshold
+ case ONE_TO_MANY:
+
+ if (getConsumer() instanceof HighlevelKafkaConsumer) {
+ throw new UnsupportedOperationException("[ONE_TO_MANY]: The high-level consumer is not supported for ONE_TO_MANY partition strategy.");
}
- newWaitingPartition.clear();
- resultPartitions = partitions;
- numPartitionsChanged = true;
- }
- break;
- // For the 1 to N mapping The initial partition number is defined by stream application
- // Afterwards, the framework will dynamically adjust the partition and allocate consumers to as less operator partitions as it can
- // and guarantee the total intake rate for each operator partition is below some threshold
- case ONE_TO_MANY:
-
- if (getConsumer() instanceof HighlevelKafkaConsumer) {
- throw new UnsupportedOperationException("[ONE_TO_MANY]: The high-level consumer is not supported for ONE_TO_MANY partition strategy.");
- }
- if (isInitialParitition || newWaitingPartition.size() != 0) {
- lastRepartitionTime = System.currentTimeMillis();
- logger.info("[ONE_TO_MANY]: Initializing partition(s)");
- // get partition metadata for topics.
- // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
- // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
- Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
-
- int size = initialPartitionCount;
- @SuppressWarnings("unchecked")
- Set<KafkaPartition>[] kps = (Set<KafkaPartition>[]) Array.newInstance((new HashSet<KafkaPartition>()).getClass(), size);
- int i = 0;
- for (Map.Entry<String, List<PartitionMetadata>> en : kafkaPartitions.entrySet()) {
- String clusterId = en.getKey();
- for (PartitionMetadata pm : en.getValue()) {
- if (kps[i % size] == null) {
- kps[i % size] = new HashSet<KafkaPartition>();
+ if (isInitialParitition || newWaitingPartition.size() != 0) {
+ lastRepartitionTime = System.currentTimeMillis();
+ logger.info("[ONE_TO_MANY]: Initializing partition(s)");
+ // get partition metadata for topics.
+ // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
+ // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
+ Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
+
+ int size = initialPartitionCount;
+ @SuppressWarnings("unchecked")
+ Set<KafkaPartition>[] kps = (Set<KafkaPartition>[])Array.newInstance((new HashSet<KafkaPartition>()).getClass(), size);
+ int i = 0;
+ for (Map.Entry<String, List<PartitionMetadata>> en : kafkaPartitions.entrySet()) {
+ String clusterId = en.getKey();
+ for (PartitionMetadata pm : en.getValue()) {
+ if (kps[i % size] == null) {
+ kps[i % size] = new HashSet<KafkaPartition>();
+ }
+ kps[i % size].add(new KafkaPartition(clusterId, consumer.topic, pm.partitionId()));
+ i++;
}
- kps[i % size].add(new KafkaPartition(clusterId, consumer.topic, pm.partitionId()));
- i++;
}
+ size = i > size ? size : i;
+ newPartitions = new ArrayList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>(size);
+ for (i = 0; i < size; i++) {
+ logger.info("[ONE_TO_MANY]: Create operator partition for kafka partition(s): {} ", StringUtils.join(kps[i], ", "));
+ newPartitions.add(createPartition(kps[i], initOffset));
+ }
+ // Add the existing partition Ids to the deleted operators
+ for (Partition<AbstractKafkaInputOperator<K>> op : partitions) {
+ deletedOperators.add(op.getPartitionedInstance().operatorId);
+ }
+ newWaitingPartition.clear();
+ resultPartitions = newPartitions;
+ numPartitionsChanged = true;
}
- size = i > size ? size : i;
- newPartitions = new ArrayList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>(size);
- for (i = 0; i < size; i++) {
- logger.info("[ONE_TO_MANY]: Create operator partition for kafka partition(s): {} ", StringUtils.join(kps[i], ", "));
- newPartitions.add(createPartition(kps[i], initOffset));
- }
- // Add the existing partition Ids to the deleted operators
- for (Partition<AbstractKafkaInputOperator<K>> op : partitions)
- {
- deletedOperators.add(op.getPartitionedInstance().operatorId);
- }
-
- newWaitingPartition.clear();
- resultPartitions = newPartitions;
- numPartitionsChanged = true;
- }
- break;
+ break;
- case ONE_TO_MANY_HEURISTIC:
- throw new UnsupportedOperationException("[ONE_TO_MANY_HEURISTIC]: Not implemented yet");
- default:
- break;
+ case ONE_TO_MANY_HEURISTIC:
+ throw new UnsupportedOperationException("[ONE_TO_MANY_HEURISTIC]: Not implemented yet");
+ default:
+ break;
}
if (numPartitionsChanged) {
@@ -689,7 +688,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
List<KafkaConsumer.KafkaMeterStats> kmsList = new LinkedList<KafkaConsumer.KafkaMeterStats>();
for (Stats.OperatorStats os : stats.getLastWindowedStats()) {
if (os != null && os.counters instanceof KafkaConsumer.KafkaMeterStats) {
- kmsList.add((KafkaConsumer.KafkaMeterStats) os.counters);
+ kmsList.add((KafkaConsumer.KafkaMeterStats)os.counters);
}
}
return kmsList;
@@ -719,12 +718,12 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
return false;
}
- if(repartitionInterval < 0){
+ if (repartitionInterval < 0) {
// if repartition is disabled
return false;
}
- if(t - lastRepartitionTime < repartitionInterval) {
+ if (t - lastRepartitionTime < repartitionInterval) {
// return false if it's still within repartitionInterval since last (re)partition
return false;
}
@@ -747,18 +746,18 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
}
Map<String, List<PartitionMetadata>> partitionsMeta = KafkaMetadataUtil.getPartitionsForTopic(consumer.brokers, consumer.getTopic());
- if(partitionsMeta == null){
+ if (partitionsMeta == null) {
//broker(s) has temporary issue to get metadata
return false;
}
for (Map.Entry<String, List<PartitionMetadata>> en : partitionsMeta.entrySet()) {
- if(en.getValue() == null){
+ if (en.getValue() == null) {
//broker(s) has temporary issue to get metadata
continue;
}
for (PartitionMetadata pm : en.getValue()) {
KafkaPartition pa = new KafkaPartition(en.getKey(), consumer.topic, pm.partitionId());
- if(!existingIds.contains(pa)){
+ if (!existingIds.contains(pa)) {
newWaitingPartition.add(pa);
}
}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java
index f0835c4..624c955 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java
@@ -16,18 +16,19 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
+
+import java.util.Properties;
+import javax.validation.constraints.NotNull;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.commons.lang3.StringUtils;
import com.datatorrent.api.Context.OperatorContext;
import com.datatorrent.api.Operator;
import kafka.javaapi.producer.Producer;
import kafka.producer.ProducerConfig;
-import org.apache.commons.lang3.StringUtils;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import javax.validation.constraints.NotNull;
-import java.util.Properties;
/**
* This is the base implementation of a Kafka output operator, which writes data to the Kafka message bus.
@@ -86,7 +87,8 @@ public abstract class AbstractKafkaOutputOperator<K, V> implements Operator
* setup producer configuration.
* @return ProducerConfig
*/
- protected ProducerConfig createKafkaProducerConfig(){
+ protected ProducerConfig createKafkaProducerConfig()
+ {
Properties prop = new Properties();
for (String propString : producerProperties.split(",")) {
if (!propString.contains("=")) {
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java
index 96dd599..f656807 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
import com.datatorrent.api.DefaultOutputPort;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java
index 85cee56..5531285 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java
@@ -16,7 +16,7 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
import java.util.HashMap;
import java.util.List;
@@ -30,8 +30,6 @@ import java.util.concurrent.Executors;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-import com.google.common.base.Joiner;
-
import kafka.consumer.ConsumerConfig;
import kafka.consumer.ConsumerIterator;
import kafka.consumer.KafkaStream;
@@ -158,8 +156,8 @@ public class HighlevelKafkaConsumer extends KafkaConsumer
Map<String, List<KafkaStream<byte[], byte[]>>> consumerMap = standardConsumer.get(e.getKey()).createMessageStreams(topicCountMap);
for (final KafkaStream<byte[], byte[]> stream : consumerMap.get(topic)) {
- consumerThreadExecutor.submit(new Runnable() {
-
+ consumerThreadExecutor.submit(new Runnable()
+ {
KafkaPartition kp = new KafkaPartition(e.getKey(), topic, -1);
public void run()
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java
index a67ff48..f59f2c8 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java
@@ -16,21 +16,8 @@
* specific language governing permissions and limitations
* under the License.
*/
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
-import com.datatorrent.api.Context;
-import com.esotericsoftware.kryo.serializers.FieldSerializer.Bind;
-import com.esotericsoftware.kryo.serializers.JavaSerializer;
-import com.google.common.collect.HashMultimap;
-import com.google.common.collect.Maps;
-import com.google.common.collect.SetMultimap;
-import kafka.message.Message;
-import org.apache.commons.io.IOUtils;
-import org.apache.commons.lang3.StringUtils;
-
-import javax.validation.constraints.NotNull;
-import javax.validation.constraints.Pattern;
-import javax.validation.constraints.Pattern.Flag;
import java.io.Closeable;
import java.io.Serializable;
import java.util.Collection;
@@ -44,6 +31,19 @@ import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.Executors;
import java.util.concurrent.ScheduledExecutorService;
import java.util.concurrent.TimeUnit;
+import javax.validation.constraints.NotNull;
+import javax.validation.constraints.Pattern;
+import javax.validation.constraints.Pattern.Flag;
+
+import org.apache.commons.io.IOUtils;
+import org.apache.commons.lang3.StringUtils;
+import com.esotericsoftware.kryo.serializers.FieldSerializer.Bind;
+import com.esotericsoftware.kryo.serializers.JavaSerializer;
+import com.google.common.collect.HashMultimap;
+import com.google.common.collect.Maps;
+import com.google.common.collect.SetMultimap;
+import com.datatorrent.api.Context;
+import kafka.message.Message;
/**
* Base Kafka Consumer class used by kafka input operator
@@ -52,9 +52,9 @@ import java.util.concurrent.TimeUnit;
*/
public abstract class KafkaConsumer implements Closeable
{
- protected final static String HIGHLEVEL_CONSUMER_ID_SUFFIX = "_stream_";
+ protected static final String HIGHLEVEL_CONSUMER_ID_SUFFIX = "_stream_";
- protected final static String SIMPLE_CONSUMER_ID_SUFFIX = "_partition_";
+ protected static final String SIMPLE_CONSUMER_ID_SUFFIX = "_partition_";
private String zookeeper;
public KafkaConsumer()
@@ -111,7 +111,7 @@ public abstract class KafkaConsumer implements Closeable
* This setting is case_insensitive
* By default it always consume from the beginning of the queue
*/
- @Pattern(flags={Flag.CASE_INSENSITIVE}, regexp = "earliest|latest")
+ @Pattern(flags = {Flag.CASE_INSENSITIVE}, regexp = "earliest|latest")
protected String initialOffset = "latest";
@@ -122,17 +122,18 @@ public abstract class KafkaConsumer implements Closeable
/**
* This method is called in setup method of the operator
*/
- public void create(){
+ public void create()
+ {
initBrokers();
holdingBuffer = new ArrayBlockingQueue<KafkaMessage>(cacheSize);
}
public void initBrokers()
{
- if(brokers!=null){
- return ;
+ if (brokers != null) {
+ return;
}
- if(zookeeperMap !=null){
+ if (zookeeperMap != null) {
brokers = HashMultimap.create();
for (String clusterId: zookeeperMap.keySet()) {
try {
@@ -158,12 +159,13 @@ public abstract class KafkaConsumer implements Closeable
/**
* The method is called in the deactivate method of the operator
*/
- public void stop() {
+ public void stop()
+ {
isAlive = false;
statsSnapShot.stop();
holdingBuffer.clear();
IOUtils.closeQuietly(this);
- };
+ }
/**
* This method is called in teardown method of the operator
@@ -227,7 +229,8 @@ public abstract class KafkaConsumer implements Closeable
}
- final protected void putMessage(KafkaPartition partition, Message msg, long offset) throws InterruptedException{
+ protected final void putMessage(KafkaPartition partition, Message msg, long offset) throws InterruptedException
+ {
// block from receiving more message
holdingBuffer.put(new KafkaMessage(partition, msg, offset));
statsSnapShot.mark(partition, msg.payloadSize());
@@ -300,7 +303,8 @@ public abstract class KafkaConsumer implements Closeable
totalBytesPerSec = _1minAvg[1];
}
- public void updateOffsets(Map<KafkaPartition, Long> offsets){
+ public void updateOffsets(Map<KafkaPartition, Long> offsets)
+ {
for (Entry<KafkaPartition, Long> os : offsets.entrySet()) {
PartitionStats ps = putPartitionStatsIfNotPresent(os.getKey());
ps.offset = os.getValue();
@@ -325,7 +329,8 @@ public abstract class KafkaConsumer implements Closeable
ps.brokerId = brokerId;
}
- private synchronized PartitionStats putPartitionStatsIfNotPresent(KafkaPartition kp){
+ private synchronized PartitionStats putPartitionStatsIfNotPresent(KafkaPartition kp)
+ {
PartitionStats ps = partitionStats.get(kp);
if (ps == null) {
@@ -347,6 +352,7 @@ public abstract class KafkaConsumer implements Closeable
this.msg = msg;
this.offSet = offset;
}
+
public KafkaPartition getKafkaPart()
{
return kafkaPart;
@@ -363,8 +369,8 @@ public abstract class KafkaConsumer implements Closeable
}
}
- public static class KafkaMeterStatsUtil {
-
+ public static class KafkaMeterStatsUtil
+ {
public static Map<KafkaPartition, Long> getOffsetsForPartitions(List<KafkaMeterStats> kafkaMeterStats)
{
Map<KafkaPartition, Long> result = Maps.newHashMap();
@@ -387,11 +393,8 @@ public abstract class KafkaConsumer implements Closeable
}
- public static class KafkaMeterStatsAggregator implements Context.CountersAggregator, Serializable{
-
- /**
- *
- */
+ public static class KafkaMeterStatsAggregator implements Context.CountersAggregator, Serializable
+ {
private static final long serialVersionUID = 729987800215151678L;
@Override
@@ -399,7 +402,7 @@ public abstract class KafkaConsumer implements Closeable
{
KafkaMeterStats kms = new KafkaMeterStats();
for (Object o : countersList) {
- if (o instanceof KafkaMeterStats){
+ if (o instanceof KafkaMeterStats) {
KafkaMeterStats subKMS = (KafkaMeterStats)o;
kms.partitionStats.putAll(subKMS.partitionStats);
kms.totalBytesPerSec += subKMS.totalBytesPerSec;
@@ -411,12 +414,8 @@ public abstract class KafkaConsumer implements Closeable
}
- public static class PartitionStats implements Serializable {
-
-
- /**
- *
- */
+ public static class PartitionStats implements Serializable
+ {
private static final long serialVersionUID = -6572690643487689766L;
public int brokerId = -1;
@@ -431,13 +430,11 @@ public abstract class KafkaConsumer implements Closeable
}
-
-
/**
* A snapshot of consuming rate within 1 min
*/
- static class SnapShot {
-
+ static class SnapShot
+ {
// msgs/s and bytes/s for each partition
/**
@@ -485,35 +482,41 @@ public abstract class KafkaConsumer implements Closeable
}
- public void start(){
- if(service==null){
+ public void start()
+ {
+ if (service == null) {
service = Executors.newScheduledThreadPool(1);
}
- service.scheduleAtFixedRate(new Runnable() {
+ service.scheduleAtFixedRate(new Runnable()
+ {
@Override
public void run()
{
moveNext();
- if(last<60)last++;
+ if (last < 60) {
+ last++;
+ }
}
}, 1, 1, TimeUnit.SECONDS);
}
- public void stop(){
- if(service!=null){
+ public void stop()
+ {
+ if (service != null) {
service.shutdown();
}
}
- public synchronized void mark(KafkaPartition partition, long bytes){
+ public synchronized void mark(KafkaPartition partition, long bytes)
+ {
msgSec[cursor]++;
msgSec[60]++;
bytesSec[cursor] += bytes;
bytesSec[60] += bytes;
... 37109 lines suppressed ...
--
To stop receiving notification emails like this one, please contact
"commits@apex.apache.org" <co...@apex.apache.org>.