You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@apex.apache.org by vr...@apache.org on 2017/12/28 17:54:01 UTC

[apex-malhar] 02/02: APEXMALHAR-2517 imports/checkstyle #662

This is an automated email from the ASF dual-hosted git repository.

vrozov pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/apex-malhar.git

commit 0d98d05ef670dcf8359681b384de1d78734f6657
Author: Apex Dev <de...@apex.apache.org>
AuthorDate: Mon Aug 14 06:53:10 2017 -0700

    APEXMALHAR-2517 imports/checkstyle #662
---
 apps/filecopy/pom.xml                              |   4 +-
 .../com/datatorrent/apps/copy/HDFSFileCopyApp.java |   4 +-
 apps/logstream/pom.xml                             |   2 +-
 .../logstream/AggregationsToRedisOperator.java     |   4 +-
 .../datatorrent/apps/logstream/Application.java    |  29 +--
 .../datatorrent/apps/logstream/Application1.java   |   5 +-
 .../apps/logstream/DimensionOperator.java          |   5 +-
 .../apps/logstream/DimensionOperatorUnifier.java   |   3 +-
 .../apps/logstream/LogScoreOperator.java           |   4 +-
 .../datatorrent/apps/logstream/LogstreamTopN.java  |   6 +-
 .../logstream/LogstreamWidgetOutputOperator.java   |   5 +-
 .../apps/logstream/RabbitMQLogsInputOperator.java  |   3 +-
 .../apps/logstream/DimensionOperatorTest.java      |   5 +-
 .../logstream/DimensionOperatorUnifierTest.java    |   5 +-
 .../apps/logstream/FilterOperatorTest.java         |   2 +-
 .../apps/logstream/LogstreamTopNTest.java          |   5 +-
 apps/pom.xml                                       |   2 +-
 benchmark/pom.xml                                  |   2 +-
 .../apache/apex/benchmark/ApplicationFixed.java    |   2 +-
 .../java/org/apache/apex/benchmark/Benchmark.java  |   2 +-
 .../apache/apex/benchmark/CouchBaseAppInput.java   |   2 +-
 .../apache/apex/benchmark/CouchBaseAppOutput.java  |   4 +-
 .../apex/benchmark/CouchBaseInputOperator.java     |   6 +-
 .../apex/benchmark/CouchBaseOutputOperator.java    |   4 +-
 .../apex/benchmark/FixedTuplesInputOperator.java   |   2 +-
 .../org/apache/apex/benchmark/RandomMapOutput.java |   2 +-
 .../apex/benchmark/RandomWordInputModule.java      |   2 +-
 .../apache/apex/benchmark/WordCountOperator.java   |   2 +-
 .../AerospikeOutputBenchmarkApplication.java       |   6 +-
 .../aerospike/AerospikeOutputOperator.java         |   6 +-
 .../algo/UniqueValueCountBenchmarkApplication.java |  14 +-
 .../CassandraOutputBenchmarkApplication.java       |   8 +-
 .../cassandra/CassandraOutputOperator.java         |   6 +-
 .../apex/benchmark/fs/FSByteOutputOperator.java    |   4 +-
 .../benchmark/fs/FSOutputOperatorBenchmark.java    |   8 +-
 .../benchmark/hive/HiveInsertBenchmarkingApp.java  |   8 +-
 .../hive/HiveMapInsertBenchmarkingApp.java         |  13 +-
 .../kafka/BenchmarkKafkaInputOperator.java         |   5 +-
 .../BenchmarkPartitionableKafkaOutputOperator.java |   2 +-
 .../apex/benchmark/kafka/KafkaInputBenchmark.java  |   9 +-
 .../apex/benchmark/kafka/KafkaOutputBenchmark.java |   2 +-
 .../apex/benchmark/kafka/KafkaTestPartitioner.java |   2 +-
 .../org/apache/apex/benchmark/package-info.java    |   2 +-
 .../script/RubyOperatorBenchmarkApplication.java   |  12 +-
 .../benchmark/spillable/SpillableBenchmarkApp.java |   4 +-
 .../spillable/SpillableTestInputOperator.java      |   2 +-
 .../benchmark/spillable/SpillableTestOperator.java |   2 +-
 .../benchmark/state/ManagedStateBenchmarkApp.java  |   6 +-
 .../apache/apex/benchmark/state/StoreOperator.java |   4 +-
 .../benchmark/stream/DevNullCounterBenchmark.java  |   6 +-
 .../apex/benchmark/stream/IntegerOperator.java     |   2 +-
 .../apex/benchmark/stream/StreamDuplicaterApp.java |   6 +-
 .../apex/benchmark/stream/StreamMergeApp.java      |   6 +-
 .../benchmark/testbench/EventClassifierApp.java    |   6 +-
 .../EventClassifierNumberToHashDoubleApp.java      |   8 +-
 .../benchmark/testbench/EventGeneratorApp.java     |   6 +-
 .../benchmark/testbench/EventIncrementerApp.java   |   6 +-
 .../benchmark/testbench/FilterClassifierApp.java   |   6 +-
 .../testbench/FilteredEventClassifierApp.java      |   6 +-
 .../apex/benchmark/testbench/HashMapOperator.java  |   5 +-
 .../testbench/RandomEventGeneratorApp.java         |   6 +-
 .../benchmark/testbench/SeedEventGeneratorApp.java |   9 +-
 .../benchmark/testbench/ThroughputCounterApp.java  |   6 +-
 .../apex/benchmark/window/AbstractGenerator.java   |   2 +-
 .../AbstractWindowedOperatorBenchmarkApp.java      |   8 +-
 .../window/KeyedWindowedOperatorBenchmarkApp.java  |   6 +-
 .../window/WindowedOperatorBenchmarkApp.java       |   2 +-
 .../apex/benchmark/ApplicationFixedTest.java       |   2 +-
 .../org/apache/apex/benchmark/BenchmarkTest.java   |   2 +-
 .../apex/benchmark/CouchBaseBenchmarkTest.java     |   2 +-
 .../apex/benchmark/accumulo/AccumuloApp.java       |  10 +-
 .../apex/benchmark/accumulo/AccumuloAppTest.java   |   2 +-
 .../aerospike/AerospikeBenchmarkAppTest.java       |   2 +-
 .../algo/UniqueValueCountBenchmarkTest.java        |   2 +-
 .../cassandra/CassandraApplicatonTest.java         |   2 +-
 .../apex/benchmark/hbase/HBaseApplicationTest.java |   2 +-
 .../hbase/HBaseCsvMappingApplication.java          |   6 +-
 .../benchmark/hive/HiveInsertBenchmarkTest.java    |   2 +-
 .../apex/benchmark/hive/HiveMapBenchmarkTest.java  |   2 +-
 .../benchmark/kafka/KafkaInputBenchmarkTest.java   |   2 +-
 .../benchmark/kafka/KafkaOutputBenchmarkTest.java  |   2 +-
 .../benchmark/memsql/MemsqlInputBenchmark.java     |   6 +-
 .../benchmark/memsql/MemsqlInputBenchmarkTest.java |  16 +-
 .../benchmark/memsql/MemsqlOutputBenchmark.java    |   6 +-
 .../memsql/MemsqlOutputBenchmarkTest.java          |   6 +-
 .../script/RubyOperatorBenchmarkAppTest.java       |   2 +-
 .../spillable/SpillableBenchmarkAppTester.java     |   2 +-
 .../spillable/SpillableDSBenchmarkTest.java        |   5 +-
 .../state/ManagedStateBenchmarkAppTest.java        |   4 +-
 .../testbench/EventClassifierAppTest.java          |   2 +-
 .../EventClassifierNumberToHashDoubleAppTest.java  |   2 +-
 .../benchmark/testbench/EventGeneratorAppTest.java |   2 +-
 .../testbench/EventIncrementerAppTest.java         |   2 +-
 .../testbench/FilterClassifierAppTest.java         |   2 +-
 .../testbench/FilteredEventClassifierAppTest.java  |   2 +-
 .../testbench/ThroughputCounterAppTest.java        |   2 +-
 .../util/serde/GenericSerdePerformanceTest.java    |   2 +-
 .../KeyedWindowedOperatorBenchmarkAppTest.java     |   2 +-
 .../window/WindowedOperatorBenchmarkAppTest.java   |   2 +-
 contrib/pom.xml                                    |  11 +-
 .../accumulo/AbstractAccumuloInputOperator.java    |  13 +-
 .../accumulo/AbstractAccumuloOutputOperator.java   |  26 +-
 .../malhar/contrib/accumulo/AccumuloStore.java     |  70 ++++--
 .../contrib/accumulo/AccumuloWindowStore.java      |  71 +++---
 .../apex/malhar/contrib/accumulo/package-info.java |   2 +-
 .../aerospike/AbstractAerospikeGetOperator.java    |  32 +--
 ...stractAerospikeNonTransactionalPutOperator.java |  30 +--
 .../AbstractAerospikeTransactionalPutOperator.java |  33 +--
 .../AerospikePOJONonTransactionalPutOperator.java  |  20 +-
 .../AerospikePOJOTransactionalPutOperator.java     |  24 +-
 .../malhar/contrib/aerospike/AerospikeStore.java   |  48 ++--
 .../aerospike/AerospikeTransactionalStore.java     |  87 ++++---
 .../malhar/contrib/aerospike/package-info.java     |   2 +-
 .../contrib/apachelog/ApacheLogInputGenerator.java |  44 ++--
 .../contrib/apachelog/ApplicationLocalLog.java     |  11 +-
 .../contrib/apachelog/ApplicationLogGenerator.java |  14 +-
 .../malhar/contrib/apachelog/GeoIPExtractor.java   |  18 +-
 .../contrib/apachelog/TimestampExtractor.java      |  11 +-
 .../contrib/apachelog/UserAgentExtractor.java      |  10 +-
 .../malhar/contrib/apachelog/package-info.java     |   2 +-
 .../malhar/contrib/apachelog/zmq/Application.java  |  16 +-
 .../malhar/contrib/avro/AvroFileInputOperator.java |   4 +-
 .../malhar}/contrib/avro/AvroFileToPojoModule.java |   2 +-
 .../apex/malhar/contrib/avro/AvroRecordHelper.java |   2 +-
 .../apex/malhar/contrib/avro/AvroToPojo.java       |  10 +-
 .../apex/malhar/contrib/avro/PojoToAvro.java       |   9 +-
 .../cassandra/AbstractCassandraInputOperator.java  |  14 +-
 ...ractCassandraTransactionableOutputOperator.java |   8 +-
 .../cassandra/AbstractUpsertOutputOperator.java    |   4 +-
 .../cassandra/CassandraPOJOInputOperator.java      |  25 +-
 .../cassandra/CassandraPOJOOutputOperator.java     |  47 ++--
 .../contrib/cassandra/CassandraPojoUtils.java      |   5 +-
 .../CassandraPreparedStatementGenerator.java       |   2 +-
 .../malhar/contrib/cassandra/CassandraStore.java   |  39 +--
 .../cassandra/CassandraTransactionalStore.java     |  45 ++--
 .../contrib/cassandra/ConnectionStateManager.java  |   4 +-
 .../contrib/cassandra/UpsertExecutionContext.java  |   4 +-
 .../malhar/contrib/cassandra/package-info.java     |   2 +-
 .../couchbase/AbstractCouchBaseInputOperator.java  |  34 ++-
 .../couchbase/AbstractCouchBaseOutputOperator.java |  29 ++-
 .../couchbase/AbstractCouchBaseSetOperator.java    |   2 +-
 .../contrib/couchbase/CouchBaseJSONSerializer.java |   7 +-
 .../couchbase/CouchBasePOJOInputOperator.java      |  20 +-
 .../contrib/couchbase/CouchBaseSerializer.java     |   2 +-
 .../malhar/contrib/couchbase/CouchBaseStore.java   |  34 ++-
 .../contrib/couchbase/CouchBaseWindowStore.java    |  33 ++-
 .../couchbase/CouchbasePOJOSetOperator.java        |   8 +-
 .../malhar/contrib/couchbase/package-info.java     |   2 +-
 .../couchdb/AbstractCouchDBInputOperator.java      |  10 +-
 .../couchdb/AbstractCouchDBOutputOperator.java     |   5 +-
 .../couchdb/AbstractMapBasedInputOperator.java     |   8 +-
 .../contrib/couchdb/CouchDBPOJOInputOperator.java  |  45 ++--
 .../contrib/couchdb/CouchDBPOJOOutputOperator.java |   7 +-
 .../apex/malhar/contrib/couchdb/CouchDbStore.java  |  10 +-
 .../couchdb/MapBasedCouchDbOutputOperator.java     |   4 +-
 .../apex/malhar/contrib/couchdb/package-info.java  |   2 +-
 .../AbstractElasticSearchInputOperator.java        |   7 +-
 .../AbstractElasticSearchOutputOperator.java       |  30 +--
 .../elasticsearch/ElasticSearchConnectable.java    |  10 +-
 .../ElasticSearchMapInputOperator.java             |   8 +-
 .../ElasticSearchMapOutputOperator.java            |  10 +-
 .../ElasticSearchPercolatorOperator.java           |  10 +-
 .../ElasticSearchPercolatorStore.java              |   5 +-
 .../malhar/contrib/elasticsearch/package-info.java |   2 +-
 .../malhar/contrib/enrich/AbstractEnricher.java    |  12 +-
 .../apex/malhar/contrib/enrich/BackendLoader.java  |   7 +-
 .../malhar/contrib/enrich/DelimitedFSLoader.java   |  11 +-
 .../apex/malhar/contrib/enrich/FSLoader.java       |   6 +-
 .../malhar/contrib/enrich/FixedWidthFSLoader.java  |   7 +-
 .../apex/malhar/contrib/enrich/JDBCLoader.java     |   6 +-
 .../apex/malhar/contrib/enrich/JsonFSLoader.java   |   2 +-
 .../apex/malhar/contrib/enrich/MapEnricher.java    |   6 +-
 .../contrib/enrich/NullValuesCacheManager.java     |   5 +-
 .../apex/malhar/contrib/enrich/POJOEnricher.java   |  13 +-
 .../apex/malhar/contrib/enrich/ReadOnlyBackup.java |   5 +-
 .../malhar/contrib/formatter/CsvFormatter.java     |  11 +-
 .../contrib/geode/AbstractGeodeInputOperator.java  |   4 +-
 .../contrib/geode/AbstractGeodeOutputOperator.java |   4 +-
 .../malhar/contrib/geode/GeodeCheckpointStore.java |  27 ++-
 .../contrib/geode/GeodeKeyValueStorageAgent.java   |   5 +-
 .../contrib/geode/GeodePOJOOutputOperator.java     |  11 +-
 .../apex/malhar/contrib/geode/GeodeStore.java      |  10 +-
 .../malhar/contrib/geode/RegionCreateFunction.java |   2 +-
 .../hbase/AbstractHBaseAppendOutputOperator.java   |  13 +-
 .../contrib/hbase/AbstractHBaseOutputOperator.java |  14 +-
 .../hbase/AbstractHBasePutOutputOperator.java      |   8 +-
 .../AbstractHBaseWindowAppendOutputOperator.java   |  28 ++-
 .../hbase/AbstractHBaseWindowOutputOperator.java   |   6 +-
 .../AbstractHBaseWindowPutOutputOperator.java      |  31 +--
 .../contrib/hbase/HBaseCsvMappingPutOperator.java  |  44 ++--
 .../apex/malhar/contrib/hbase/HBaseFieldInfo.java  | 133 +++++-----
 .../contrib/hbase/HBaseFieldValueGenerator.java    |   6 +-
 .../malhar/contrib/hbase/HBaseGetOperator.java     |   2 +-
 .../malhar/contrib/hbase/HBaseInputOperator.java   |   5 +-
 .../hbase/HBaseNameValueCsvPutOperator.java        |  71 +++---
 .../malhar/contrib/hbase/HBaseOperatorBase.java    |  11 +-
 .../contrib/hbase/HBasePOJOInputOperator.java      |  28 +--
 .../malhar/contrib/hbase/HBasePOJOPutOperator.java |  13 +-
 .../malhar/contrib/hbase/HBaseScanOperator.java    |  32 +--
 .../apex/malhar/contrib/hbase/HBaseStore.java      |  80 +++---
 .../malhar/contrib/hbase/HBaseWindowStore.java     |  55 +++--
 .../apex/malhar}/contrib/hbase/OutputAdapter.java  |   4 +-
 .../apex/malhar/contrib/hbase/package-info.java    |   2 +-
 .../apex/malhar/contrib/jython/PythonOperator.java |  18 +-
 .../apex/malhar/contrib/jython/package-info.java   |   2 +-
 .../AbstractExactlyOnceKafkaOutputOperator.java    |  36 ++-
 .../contrib/kafka/AbstractKafkaInputOperator.java  | 269 ++++++++++-----------
 .../contrib/kafka/AbstractKafkaOutputOperator.java |  18 +-
 .../AbstractKafkaSinglePortInputOperator.java      |   2 +-
 .../contrib/kafka/HighlevelKafkaConsumer.java      |   8 +-
 .../apex/malhar/contrib/kafka/KafkaConsumer.java   | 116 ++++-----
 .../malhar/contrib/kafka/KafkaJsonEncoder.java     |   2 +-
 .../malhar/contrib/kafka/KafkaMetadataUtil.java    |  22 +-
 .../apex/malhar/contrib/kafka/KafkaPartition.java  |  32 ++-
 .../KafkaSinglePortByteArrayInputOperator.java     |  32 ++-
 .../kafka/KafkaSinglePortOutputOperator.java       |   2 +-
 .../kafka/KafkaSinglePortStringInputOperator.java  |   8 +-
 .../apex/malhar/contrib/kafka/OffsetManager.java   |   2 +-
 .../contrib/kafka/POJOKafkaOutputOperator.java     |   8 +-
 .../malhar/contrib/kafka/SimpleKafkaConsumer.java  | 114 ++++-----
 .../apex/malhar/contrib/kafka/package-info.java    |   2 +-
 .../kinesis/AbstractKinesisInputOperator.java      | 194 +++++++--------
 .../kinesis/AbstractKinesisOutputOperator.java     |  43 ++--
 .../kinesis/KinesisByteArrayInputOperator.java     |   5 +-
 .../kinesis/KinesisByteArrayOutputOperator.java    |   2 +-
 .../malhar/contrib/kinesis/KinesisConsumer.java    |  62 +++--
 .../kinesis/KinesisStringInputOperator.java        |   8 +-
 .../kinesis/KinesisStringOutputOperator.java       |   3 +-
 .../apex/malhar/contrib/kinesis/KinesisUtil.java   |  41 ++--
 .../apex/malhar/contrib/kinesis/ShardManager.java  |   2 +-
 .../apex/malhar/contrib/kinesis/package-info.java  |   2 +-
 .../memcache/AbstractMemcacheInputOperator.java    |   4 +-
 .../memcache/AbstractMemcacheOutputOperator.java   |   7 +-
 .../memcache/MemcachePOJOOutputOperator.java       |  18 +-
 .../malhar/contrib/memcache/MemcacheStore.java     |  16 +-
 .../apex/malhar/contrib/memcache/package-info.java |   2 +-
 .../AbstractMemcacheInputOperator.java             |   4 +-
 .../AbstractMemcacheOutputOperator.java            |   7 +-
 .../contrib/memcache_whalin/MemcacheStore.java     |  14 +-
 .../contrib/memcache_whalin/package-info.java      |   2 +-
 .../memsql/AbstractMemsqlInputOperator.java        |   4 +-
 .../memsql/AbstractMemsqlOutputOperator.java       |   4 +-
 .../contrib/memsql/MemsqlPOJOInputOperator.java    |  98 +++-----
 .../contrib/memsql/MemsqlPOJOOutputOperator.java   |  57 ++---
 .../apex/malhar/contrib/memsql/MemsqlStore.java    |   4 +-
 .../apex/malhar/contrib/memsql/package-info.java   |   2 +-
 .../misc/algo/AbstractStreamPatternMatcher.java    |   4 +-
 .../malhar/contrib/misc/algo/AllAfterMatchMap.java |   4 +-
 .../apex/malhar/contrib/misc/algo/DistinctMap.java |   6 +-
 .../malhar/contrib/misc/algo/FilterKeyVals.java    |   4 +-
 .../contrib/misc/algo/FilterKeysHashMap.java       |   4 +-
 .../malhar/contrib/misc/algo/FilterKeysMap.java    |   6 +-
 .../malhar/contrib/misc/algo/FirstMatchMap.java    |   4 +-
 .../apex/malhar/contrib/misc/algo/FirstN.java      |   3 +-
 .../malhar/contrib/misc/algo/FirstTillMatch.java   |   4 +-
 .../malhar/contrib/misc/algo/InsertSortDesc.java   |   5 +-
 .../apex/malhar/contrib/misc/algo/InvertIndex.java |   4 +-
 .../malhar/contrib/misc/algo/InvertIndexArray.java |   4 +-
 .../malhar/contrib/misc/algo/LastMatchMap.java     |   4 +-
 .../contrib/misc/algo/LeastFrequentKeyMap.java     |   8 +-
 .../misc/algo/LeastFrequentKeyValueMap.java        |   4 +-
 .../contrib/misc/algo/MostFrequentKeyMap.java      |   8 +-
 .../contrib/misc/algo/MostFrequentKeyValueMap.java |   4 +-
 .../apex/malhar/contrib/misc/algo/Sampler.java     |   4 +-
 .../{aerospike => misc/algo}/package-info.java     |   2 +-
 .../apex/malhar/contrib/misc/math/Change.java      |   3 +-
 .../apex/malhar/contrib/misc/math/ChangeAlert.java |   5 +-
 .../contrib/misc/math/ChangeAlertKeyVal.java       |   4 +-
 .../malhar/contrib/misc/math/ChangeAlertMap.java   |   2 +-
 .../malhar/contrib/misc/math/ChangeKeyVal.java     |   5 +-
 .../malhar/contrib/misc/math/CompareExceptMap.java |   5 +-
 .../apex/malhar/contrib/misc/math/CompareMap.java  |   3 +-
 .../apex/malhar/contrib/misc/math/CountKeyVal.java |   6 +-
 .../apex/malhar/contrib/misc/math/ExceptMap.java   |   5 +-
 .../apex/malhar/contrib/misc/math/Quotient.java    |   3 +-
 .../apex/malhar/contrib/misc/math/QuotientMap.java |   2 +-
 .../apex/malhar/contrib/misc/math/SumCountMap.java |   6 +-
 .../{aerospike => misc/math}/package-info.java     |   2 +-
 .../contrib/misc/streamquery/DeleteOperator.java   |   3 +-
 .../misc/streamquery/GroupByHavingOperator.java    |   7 +-
 .../misc/streamquery/InnerJoinOperator.java        |   5 +-
 .../contrib/misc/streamquery/SelectOperator.java   |   5 +-
 .../contrib/misc/streamquery/UpdateOperator.java   |   3 +-
 .../streamquery/condition/BetweenCondition.java    |   2 +-
 .../streamquery/condition/CompoundCondition.java   |   2 +-
 .../streamquery/condition/EqualValueCondition.java |   2 +-
 .../misc/streamquery/condition/InCondition.java    |   2 +-
 .../misc/streamquery/condition/LikeCondition.java  |   2 +-
 .../misc/streamquery/index/BinaryExpression.java   |   2 +-
 .../contrib/misc/streamquery/index/MidIndex.java   |   2 +-
 .../misc/streamquery/index/NegateExpression.java   |   2 +-
 .../misc/streamquery/index/RoundDoubleIndex.java   |   2 +-
 .../misc/streamquery/index/StringCaseIndex.java    |   2 +-
 .../misc/streamquery/index/StringLenIndex.java     |   2 +-
 .../misc/streamquery/index/SumExpression.java      |   2 +-
 .../misc/streamquery/index/UnaryExpression.java    |   2 +-
 .../mongodb/MongoDBArrayListOutputOperator.java    |   7 +-
 .../malhar/contrib/mongodb/MongoDBConnectable.java |  13 +-
 .../mongodb/MongoDBHashMapOutputOperator.java      |   9 +-
 .../contrib/mongodb/MongoDBInputOperator.java      |  21 +-
 .../contrib/mongodb/MongoDBOutputOperator.java     |  56 ++---
 .../contrib/mongodb/MongoDBPOJOOutputOperator.java |  27 +--
 .../apex/malhar/contrib/mongodb/package-info.java  |   2 +-
 .../contrib/mqtt/AbstractMqttInputOperator.java    |  30 +--
 .../contrib/mqtt/AbstractMqttOutputOperator.java   |  13 +-
 .../mqtt/AbstractSinglePortMqttInputOperator.java  |   6 +-
 .../mqtt/AbstractSinglePortMqttOutputOperator.java |   2 +-
 .../apex/malhar/contrib/mqtt/MqttClientConfig.java |   4 +-
 .../apex/malhar/contrib/mqtt/package-info.java     |   2 +-
 .../contrib/nifi/AbstractNiFiInputOperator.java    |   5 +-
 .../contrib/nifi/AbstractNiFiOutputOperator.java   |   2 +-
 .../nifi/AbstractNiFiSinglePortInputOperator.java  |   2 +-
 .../apex/malhar/contrib/nifi/NiFiDataPacket.java   |   2 +-
 .../malhar/contrib/nifi/NiFiDataPacketBuilder.java |   2 +-
 .../contrib/nifi/NiFiSinglePortInputOperator.java  |   2 +-
 .../contrib/nifi/NiFiSinglePortOutputOperator.java |   2 +-
 .../contrib/nifi/StandardNiFiDataPacket.java       |   2 +-
 .../contrib/parquet/AbstractParquetFileReader.java |   6 +-
 .../contrib/parquet/ParquetFilePOJOReader.java     |   6 +-
 .../malhar/contrib/parser/AbstractCsvParser.java   |  66 +++--
 .../contrib/parser/CellProcessorBuilder.java       |   5 +-
 .../malhar/contrib/parser/CommonLogParser.java     |  48 ++--
 .../apex/malhar/contrib/parser/CsvParser.java      |  10 +-
 .../malhar/contrib/parser/DelimitedSchema.java     |   2 +-
 .../malhar/contrib/parser/FixedWidthParser.java    |  20 +-
 .../malhar/contrib/parser/FixedWidthSchema.java    |  17 +-
 .../apex/malhar/contrib/parser/JsonParser.java     |   6 +-
 .../apex/malhar/contrib/parser/LogParser.java      |  11 +-
 .../malhar/contrib/parser/LogSchemaDetails.java    |  18 +-
 .../apex/malhar/contrib/parser/RegexParser.java    |  10 +-
 .../apache/apex/malhar/contrib/parser/Schema.java  |   2 +-
 .../malhar/contrib/parser/StreamingJsonParser.java |  10 +-
 .../apex/malhar/contrib/parser/package-info.java   |   2 +-
 .../apex/malhar/contrib/r/REngineConnectable.java  |  12 +-
 .../org/apache/apex/malhar/contrib/r/RMax.java     |  14 +-
 .../org/apache/apex/malhar/contrib/r/RMin.java     |  14 +-
 .../org/apache/apex/malhar/contrib/r/RScript.java  |  87 +++----
 .../apex/malhar/contrib/r/RStandardDeviation.java  |  12 +-
 .../apache/apex/malhar/contrib/r/package-info.java |   2 +-
 .../rabbitmq/AbstractRabbitMQInputOperator.java    |  62 ++---
 .../rabbitmq/AbstractRabbitMQOutputOperator.java   |  48 ++--
 .../AbstractSinglePortRabbitMQInputOperator.java   |  15 +-
 .../AbstractSinglePortRabbitMQOutputOperator.java  |   5 +-
 .../contrib/rabbitmq/RabbitMQInputOperator.java    |   2 +-
 .../contrib/rabbitmq/RabbitMQOutputOperator.java   |   2 +-
 .../apex/malhar/contrib/rabbitmq/package-info.java |   2 +-
 .../AbstractRedisAggregateOutputOperator.java      |   6 +-
 .../contrib/redis/AbstractRedisInputOperator.java  |  21 +-
 .../apex/malhar/contrib/redis/LettuceStore.java    |  13 +-
 .../apex/malhar/contrib/redis/NumberSummation.java |  44 ++--
 .../redis/RedisKeyValPairOutputOperator.java       |   5 +-
 .../contrib/redis/RedisKeyValueInputOperator.java  |   4 +-
 .../redis/RedisMapAsValueInputOperator.java        |   4 +-
 .../contrib/redis/RedisMapOutputOperator.java      |   2 +-
 ...disNumberSummationKeyValPairOutputOperator.java |   5 +-
 .../RedisNumberSummationMapOutputOperator.java     |   2 +-
 .../contrib/redis/RedisPOJOInputOperator.java      |  38 +--
 .../contrib/redis/RedisPOJOOutputOperator.java     |  14 +-
 .../apex/malhar/contrib/redis/RedisStore.java      |  34 ++-
 .../apex/malhar/contrib/redis/package-info.java    |   2 +-
 .../contrib/romesyndication/RomeFeedEntry.java     |   6 +-
 .../romesyndication/RomeStreamProvider.java        |   2 +-
 .../romesyndication/RomeSyndicationOperator.java   |  30 +--
 .../contrib/romesyndication/package-info.java      |   2 +-
 .../apex/malhar/contrib/ruby/RubyOperator.java     |  58 ++---
 .../contrib/solr/AbstractSolrInputOperator.java    |   8 +-
 .../contrib/solr/AbstractSolrOutputOperator.java   |  12 +-
 .../contrib/solr/CloudSolrServerConnector.java     |   2 +-
 .../solr/ConcurrentUpdateSolrServerConnector.java  |   2 +-
 .../contrib/solr/HttpSolrServerConnector.java      |   2 +-
 .../contrib/solr/LBHttpSolrServerConnector.java    |   2 +-
 .../malhar/contrib/solr/SolrInputOperator.java     |   6 +-
 .../malhar/contrib/solr/SolrOutputOperator.java    |   2 +-
 .../malhar/contrib/solr/SolrServerConnector.java   |   5 +-
 .../apex/malhar/contrib/solr/package-info.java     |   2 +-
 .../splunk/AbstractSplunkInputOperator.java        |  24 +-
 .../contrib/splunk/SplunkInputFromForwarder.java   |  56 +++--
 .../malhar/contrib/splunk/SplunkInputOperator.java |   2 +-
 .../apex/malhar/contrib/splunk/SplunkStore.java    |  38 +--
 .../contrib/splunk/SplunkTcpOutputOperator.java    |  38 +--
 .../apex/malhar/contrib/splunk/package-info.java   |   2 +-
 .../contrib/sqlite/SqliteStreamOperator.java       |  27 +--
 .../apex/malhar/contrib/sqlite/package-info.java   |   2 +-
 .../malhar/contrib/twitter/TwitterSampleInput.java |  40 +--
 .../zmq/AbstractBaseZeroMQInputOperator.java       |  26 +-
 .../zmq/AbstractBaseZeroMQOutputOperator.java      |  10 +-
 .../zmq/AbstractSinglePortZeroMQInputOperator.java |  15 +-
 .../AbstractSinglePortZeroMQOutputOperator.java    |   2 +-
 .../SimpleSinglePortZeroMQPullInputOperator.java   |   5 +-
 ...pleSinglePortZeroMQPullStringInputOperator.java |   2 +-
 .../malhar/contrib/zmq/ZeroMQInputOperator.java    |   5 +-
 .../malhar/contrib/zmq/ZeroMQOutputOperator.java   |  15 +-
 .../apex/malhar/contrib/zmq/package-info.java      |   2 +-
 .../accumulo/AccumuloInputOperatorTest.java        |  29 ++-
 .../accumulo/AccumuloOutputOperatorTest.java       |  30 +--
 .../accumulo/AccumuloRowTupleGenerator.java        |  21 +-
 .../contrib/accumulo/AccumuloTestHelper.java       |  39 +--
 .../malhar/contrib/accumulo/AccumuloTuple.java     |  31 ++-
 .../contrib/accumulo/AccumuloTupleCollector.java   |  18 +-
 .../contrib/aerospike/AerospikeOperatorTest.java   |  93 +++----
 ...rospikePOJONonTransactionalPutOperatorTest.java |  23 +-
 .../AerospikePOJOTransactionalPutOperatorTest.java |  23 +-
 .../contrib/aerospike/AerospikeTestUtils.java      | 123 +++++-----
 .../apachelog/ApacheLogInputGeneratorTest.java     |   6 +-
 .../contrib/apachelog/ApplicationLocalLogTest.java |  18 +-
 .../apachelog/ApplicationLogGeneratorTest.java     |  18 +-
 .../contrib/apachelog/zmq/ApplicationTest.java     |  16 +-
 .../contrib/avro/AvroFileInputOperatorTest.java    |  27 ++-
 .../contrib/avro/AvroFileToPojoModuleTest.java     |   8 +-
 .../apex/malhar/contrib/avro/AvroToPojoTest.java   |   6 +-
 .../apex/malhar/contrib/avro/PojoToAvroTest.java   |   7 +-
 .../AbstractUpsertOutputOperatorCodecsTest.java    |   7 +-
 ...bstractUpsertOutputOperatorCompositePKTest.java |   7 +-
 .../AbstractUpsertOutputOperatorCountersTest.java  |   8 +-
 .../apex/malhar/contrib/cassandra/Address.java     |   2 +-
 .../malhar/contrib/cassandra/AddressCodec.java     |   2 +-
 .../CassandraLookupCacheBackedOperatorTest.java    |  16 +-
 .../contrib/cassandra/CassandraOperatorTest.java   |  33 ++-
 .../contrib/cassandra/CompositePrimaryKeyRow.java  |   2 +-
 .../CompositePrimaryKeyUpdateOperator.java         |   3 +-
 .../contrib/cassandra/CounterColumnTableEntry.java |   2 +-
 .../cassandra/CounterColumnUpdatesOperator.java    |   3 +-
 .../apex/malhar/contrib/cassandra/FullName.java    |   2 +-
 .../malhar/contrib/cassandra/FullNameCodec.java    |   2 +-
 .../malhar/contrib/cassandra/TestInputPojo.java    |   3 +-
 .../apache/apex/malhar/contrib/cassandra/User.java |   2 +-
 .../contrib/cassandra/UserUpsertOperator.java      |   3 +-
 .../malhar/contrib/couchbase/CouchBaseGetTest.java |  20 +-
 .../couchbase/CouchBaseInputOperatorTest.java      |  30 ++-
 .../couchbase/CouchBaseOutputOperatorTest.java     |  28 +--
 .../contrib/couchbase/CouchBasePOJOTest.java       |  39 ++-
 .../malhar/contrib/couchbase/CouchBaseSetTest.java |  17 +-
 .../contrib/couchbase/TestComplexPojoInput.java    |   2 +-
 .../apex/malhar/contrib/couchbase/TestPojo.java    |   3 +-
 .../contrib/couchdb/CouchDBInputOperatorTest.java  |  14 +-
 .../contrib/couchdb/CouchDBOutputOperatorTest.java |  13 +-
 .../couchdb/CouchDBPOJOInputOperatorTest.java      |   8 +-
 .../malhar/contrib/couchdb/CouchDBTestHelper.java  |   8 +-
 .../couchdb/PagedCouchDBInputOperatorTest.java     |  11 +-
 .../apex/malhar/contrib/couchdb/TestInputPOJO.java |   2 +-
 .../elasticsearch/ElasticSearchOperatorTest.java   |  16 +-
 .../elasticsearch/ElasticSearchPercolateTest.java  |  32 ++-
 .../apex/malhar/contrib/enrich/EmployeeOrder.java  |   2 +-
 .../malhar/contrib/enrich/FileEnrichmentTest.java  |   7 +-
 .../apex/malhar/contrib/enrich/JDBCLoaderTest.java |   5 +-
 .../malhar/contrib/enrich/MapEnricherTest.java     |  10 +-
 .../apache/apex/malhar/contrib/enrich/Order.java   |   2 +-
 .../malhar/contrib/enrich/POJOEnricherTest.java    |   6 +-
 .../malhar/contrib/formatter/CsvFormatterTest.java |  12 +-
 .../contrib/geode/GeodeCheckpointStoreTest.java    |   4 +-
 .../geode/GeodeKeyValueStorageAgentTest.java       |   4 +-
 .../malhar/contrib/geode/GeodeOperatorTest.java    |   4 +-
 .../contrib/geode/GeodePOJOOperatorTest.java       |   9 +-
 .../apex/malhar/contrib/geode/GeodeStoreTest.java  |   2 +-
 .../contrib/hbase/HBaseColTupleGenerator.java      |   6 +-
 .../hbase/HBaseCsvMappingPutOperatorTest.java      |  12 +-
 .../malhar/contrib/hbase/HBaseGetOperatorTest.java |   9 +-
 .../hbase/HBaseNameValueCsvPutOperatorTest.java    |   9 +-
 .../contrib/hbase/HBasePOJOInputOperatorTest.java  |  45 ++--
 .../contrib/hbase/HBasePOJOPutOperatorTest.java    | 134 +++++-----
 .../contrib/hbase/HBaseRowStringGenerator.java     |  20 +-
 .../contrib/hbase/HBaseRowTupleGenerator.java      |   6 +-
 .../contrib/hbase/HBaseScanOperatorTest.java       |  10 +-
 .../apex/malhar/contrib/hbase/HBaseTestHelper.java |  70 +++---
 .../HBaseTransactionalAppendOperatorTest.java      |  41 ++--
 .../hbase/HBaseTransactionalPutOperatorTest.java   |  79 +++---
 .../apex/malhar/contrib/hbase/HBaseTuple.java      |   2 +-
 .../malhar/contrib/hbase/HBaseTupleCollector.java  |  27 +--
 .../apex/malhar/contrib/hbase/HBaseUtil.java       |  30 +--
 .../malhar/contrib/helper/CollectorModule.java     |   8 +-
 .../contrib/helper/MessageQueueTestHelper.java     |  11 +-
 .../apex/malhar/contrib/helper/SourceModule.java   |   6 +-
 .../malhar/contrib/jython/PythonOperatorTest.java  |   6 +-
 .../kafka/KafkaExactlyOnceOutputOperatorTest.java  |  16 +-
 .../contrib/kafka/KafkaInputOperatorTest.java      |  30 +--
 .../contrib/kafka/KafkaOperatorTestBase.java       |  28 ++-
 .../contrib/kafka/KafkaOutputOperatorTest.java     |  22 +-
 .../kafka/KafkaPartitionableInputOperatorTest.java |  24 +-
 .../malhar/contrib/kafka/KafkaSimpleConsumer.java  |  11 +-
 .../malhar/contrib/kafka/KafkaTestConsumer.java    |  12 +-
 .../malhar/contrib/kafka/KafkaTestPartitioner.java |   9 +-
 .../malhar/contrib/kafka/KafkaTestProducer.java    |  25 +-
 .../malhar/contrib/kafka/OffsetManagerTest.java    |  49 ++--
 .../contrib/kafka/SimpleKakfaConsumerTest.java     |   5 +-
 .../KinesisByteArrayOutputOperatorTest.java        |  23 +-
 .../contrib/kinesis/KinesisInputOperatorTest.java  |   8 +-
 .../contrib/kinesis/KinesisOperatorTestBase.java   |  30 +--
 .../contrib/kinesis/KinesisOutputOperatorTest.java |  48 ++--
 .../KinesisPartitionableInputOperatorTest.java     |  13 +-
 .../kinesis/KinesisStringOutputOperatorTest.java   |   7 +-
 .../contrib/kinesis/KinesisTestConsumer.java       |  44 ++--
 .../contrib/kinesis/KinesisTestProducer.java       |  21 +-
 .../malhar/contrib/kinesis/ShardManagerTest.java   |  57 ++---
 .../contrib/memcache/MemcacheOperatorTest.java     |  10 +-
 .../contrib/memcache/MemcachePOJOOperatorTest.java |  24 +-
 .../MemcacheWhalinOperatorTest.java                |   4 +-
 .../memsql/AbstractMemsqlInputOperatorTest.java    |  57 +++--
 .../memsql/AbstractMemsqlOutputOperatorTest.java   |  55 ++---
 .../malhar/contrib/memsql/MemsqlInputOperator.java |  12 +-
 .../malhar/contrib/memsql/MemsqlStoreTest.java     |   2 +-
 .../apex/malhar/contrib/memsql/TestInputPojo.java  |   2 +-
 .../algo/AbstractStreamPatternMatcherTest.java     |   3 +-
 .../contrib/misc/algo/AllAfterMatchMapTest.java    |   2 +-
 .../malhar/contrib/misc/algo/DistinctMapTest.java  |   2 +-
 .../contrib/misc/algo/FilterKeyValsTest.java       |   2 +-
 .../contrib/misc/algo/FilterKeysHashMapTest.java   |   2 +-
 .../contrib/misc/algo/FilterKeysMapTest.java       |   2 +-
 .../contrib/misc/algo/FirstMatchMapTest.java       |   2 +-
 .../apex/malhar/contrib/misc/algo/FirstNTest.java  |   2 +-
 .../contrib/misc/algo/FirstTillMatchTest.java      |   2 +-
 .../contrib/misc/algo/InsertSortDescTest.java      |   2 +-
 .../contrib/misc/algo/InvertIndexArrayTest.java    |   4 +-
 .../malhar/contrib/misc/algo/InvertIndexTest.java  |   4 +-
 .../malhar/contrib/misc/algo/LastMatchMapTest.java |   2 +-
 .../contrib/misc/algo/LeastFrequentKeyMapTest.java |   2 +-
 .../misc/algo/LeastFrequentKeyValueMapTest.java    |   2 +-
 .../malhar/contrib/misc/algo/MatchMapTest.java     |   6 +-
 .../contrib/misc/algo/MostFrequentKeyMapTest.java  |   2 +-
 .../misc/algo/MostFrequentKeyValueMapTest.java     |   2 +-
 .../apex/malhar/contrib/misc/algo/SamplerTest.java |   2 +-
 .../contrib/misc/math/ChangeAlertKeyValTest.java   |   4 +-
 .../contrib/misc/math/ChangeAlertMapTest.java      |   2 +-
 .../malhar/contrib/misc/math/ChangeAlertTest.java  |   4 +-
 .../malhar/contrib/misc/math/ChangeKeyValTest.java |   4 +-
 .../apex/malhar/contrib/misc/math/ChangeTest.java  |   2 +-
 .../contrib/misc/math/CompareExceptMapTest.java    |   2 +-
 .../malhar/contrib/misc/math/CompareMapTest.java   |   2 +-
 .../malhar/contrib/misc/math/CountKeyValTest.java  |   4 +-
 .../malhar/contrib/misc/math/ExceptMapTest.java    |   2 +-
 .../malhar/contrib/misc/math/QuotientMapTest.java  |   2 +-
 .../malhar/contrib/misc/math/SumCountMapTest.java  |   2 +-
 .../misc/streamquery/DeleteOperatorTest.java       |   3 +-
 .../streamquery/FullOuterJoinOperatorTest.java     |   9 +-
 .../misc/streamquery/GroupByOperatorTest.java      |   5 +-
 .../misc/streamquery/HavingOperatorTest.java       |   5 +-
 .../misc/streamquery/InnerJoinOperatorTest.java    |  10 +-
 .../streamquery/LeftOuterJoinOperatorTest.java     |   8 +-
 .../misc/streamquery/OrderByOperatorTest.java      |   4 +-
 .../streamquery/RightOuterJoinOperatorTest.java    |   8 +-
 .../misc/streamquery/SelectOperatorTest.java       |   5 +-
 .../misc/streamquery/SelectTopOperatorTest.java    |   2 +-
 .../misc/streamquery/UpdateOperatorTest.java       |   3 +-
 .../streamquery/advanced/BetweenConditionTest.java |   5 +-
 .../advanced/CompoundConditionTest.java            |   5 +-
 .../misc/streamquery/advanced/InConditionTest.java |   5 +-
 .../streamquery/advanced/LikeConditionTest.java    |   5 +-
 .../misc/streamquery/advanced/NegateIndexTest.java |   3 +-
 .../streamquery/advanced/SelectAverageTest.java    |   3 +-
 .../misc/streamquery/advanced/SelectCountTest.java |   3 +-
 .../streamquery/advanced/SelectFirstLastTest.java  |   3 +-
 .../streamquery/advanced/SelectMaxMinTest.java     |   3 +-
 .../misc/streamquery/advanced/SumIndexTest.java    |   3 +-
 .../contrib/mongodb/MongoDBInputOperatorTest.java  |  20 +-
 .../contrib/mongodb/MongoDBOutputOperatorTest.java |  32 ++-
 .../apex/malhar/contrib/mongodb/TestPOJO.java      |   2 +-
 .../malhar/contrib/mqtt/MqttInputOperatorTest.java |  31 +--
 .../contrib/mqtt/MqttOutputOperatorTest.java       |  30 +--
 .../nifi/NiFiSinglePortInputOperatorTest.java      |  15 +-
 .../nifi/NiFiSinglePortOutputOperatorTest.java     |  29 +--
 .../nifi/demo/TestNiFiInputApplication.java        |   6 +-
 .../nifi/demo/TestNiFiOutputApplication.java       |  17 +-
 .../malhar/contrib/nifi/mock/MockDataPacket.java   |   2 +-
 .../contrib/nifi/mock/MockSiteToSiteClient.java    |   2 +-
 .../malhar/contrib/nifi/mock/MockTransaction.java  |   2 +-
 .../contrib/parquet/ParquetFilePOJOReaderTest.java |  37 ++-
 .../apex/malhar/contrib/parser/CSVParserTest.java  |  34 +--
 .../malhar/contrib/parser/CommonLogParserTest.java |   3 +-
 .../malhar/contrib/parser/CsvPOJOParserTest.java   |   8 +-
 .../apex/malhar/contrib/parser/FixedWidthTest.java |  10 +-
 .../contrib/parser/JsonParserApplicationTest.java  |   8 +-
 .../apex/malhar/contrib/parser/JsonParserTest.java |  10 +-
 .../apex/malhar/contrib/parser/LogParserTest.java  |  36 +--
 .../malhar/contrib/parser/RegexParserTest.java     |  38 ++-
 .../contrib/parser/StreamingJsonParserTest.java    |   8 +-
 .../apex/malhar/contrib/r/RMaxOperatorTest.java    |   6 +-
 .../apex/malhar/contrib/r/RMinOperatorTest.java    |   6 +-
 .../r/RScriptLinearRegressionPredictorTest.java    |  14 +-
 .../contrib/r/RScriptOperatorBooleanTest.java      |  16 +-
 .../contrib/r/RScriptOperatorDoubleArrayTest.java  |  10 +-
 .../malhar/contrib/r/RScriptOperatorIntTest.java   |   6 +-
 .../malhar/contrib/r/RScriptOperatorRealTest.java  |   8 +-
 .../malhar/contrib/r/RScriptOperatorStrTest.java   |   4 +-
 .../contrib/r/RStandardDeviationOperatorTest.java  |   8 +-
 .../rabbitmq/RabbitMQInputOperatorBenchmark.java   |   2 +-
 .../rabbitmq/RabbitMQInputOperatorTest.java        |  22 +-
 .../rabbitmq/RabbitMQOutputOperatorBenchmark.java  |   2 +-
 .../rabbitmq/RabbitMQOutputOperatorTest.java       |  14 +-
 .../contrib/redis/RedisInputOperatorTest.java      |  15 +-
 .../malhar/contrib/redis/RedisOperatorTest.java    |   4 +-
 .../contrib/redis/RedisPOJOOperatorTest.java       |  20 +-
 .../RomeSyndicationOperatorTest.java               |  25 +-
 .../apex/malhar/contrib/ruby/RubyOperatorTest.java |  18 +-
 .../splunk/SplunkInputFromForwarderTest.java       |  17 +-
 .../contrib/splunk/SplunkInputOperatorTest.java    |  11 +-
 .../splunk/SplunkTcpOutputOperatorTest.java        |  32 +--
 .../util/FieldValueSerializableGenerator.java      |  50 ++--
 .../contrib/util/POJOTupleGenerateOperator.java    |  36 ++-
 .../apache/apex/malhar/contrib/util/TestPOJO.java  |  64 ++---
 .../contrib/util/TupleCacheOutputOperator.java     |  13 +-
 .../contrib/util/TupleGenerateCacheOperator.java   |   9 +-
 .../apex/malhar/contrib/util/TupleGenerator.java   |  33 +--
 .../contrib/zmq/ZeroMQInputOperatorBenchmark.java  |   5 +-
 .../contrib/zmq/ZeroMQInputOperatorTest.java       |  21 +-
 .../malhar/contrib/zmq/ZeroMQMessageGenerator.java |  12 +-
 .../malhar/contrib/zmq/ZeroMQMessageReceiver.java  |  13 +-
 .../contrib/zmq/ZeroMQOutputOperatorBenchmark.java |   2 +-
 .../contrib/zmq/ZeroMQOutputOperatorTest.java      |  21 +-
 examples/csvformatter/pom.xml                      |   2 +-
 .../apex/examples/csvformatter/Application.java    |   6 +-
 .../examples/csvformatter/HDFSOutputOperator.java  |   3 +-
 examples/dedup/pom.xml                             |   2 +-
 .../apache/apex/examples/dedup/Application.java    |   2 +-
 examples/distributedistinct/pom.xml                |   2 +-
 .../examples/distributeddistinct/Application.java  |  10 +-
 .../distributeddistinct/CountVerifier.java         |   3 +-
 .../IntegerUniqueValueCountAppender.java           |   4 +-
 .../distributeddistinct/RandomKeyValGenerator.java |   5 +-
 .../distributeddistinct/StatefulApplication.java   |   8 +-
 .../UniqueValueCountAppender.java                  |   7 +-
 .../DistributedDistinctTest.java                   |   8 +-
 .../StatefulUniqueCountTest.java                   |   5 +-
 examples/dynamic-partition/pom.xml                 |   2 +-
 .../java/org/apache/apex/examples/dynamic/App.java |   3 +-
 examples/echoserver/pom.xml                        |   2 +-
 examples/enricher/pom.xml                          |   2 +-
 .../examples/enricher/EnricherAppWithJSONFile.java |   6 +-
 .../apex/examples/enricher/LineOutputOperator.java |   2 +-
 examples/exactly-once/pom.xml                      |   2 +-
 .../exactlyonce/ExactlyOnceFileOutputApp.java      |   6 +-
 .../exactlyonce/ExactlyOnceJdbcOutputApp.java      |  10 +-
 .../exactlyonce/ExactlyOnceJdbcOutputTest.java     |   2 +-
 examples/filter/pom.xml                            |   2 +-
 .../apache/apex/examples/filter/Application.java   |   6 +-
 examples/frauddetect/pom.xml                       |   2 +-
 .../apex/examples/frauddetect/Application.java     |  14 +-
 .../frauddetect/AverageAlertingOperator.java       |   3 +-
 .../apex/examples/frauddetect/BankIdNumberKey.java |   2 +-
 .../frauddetect/BankIdNumberSamplerOperator.java   |   4 +-
 .../CreditCardAmountSamplerOperator.java           |   2 +-
 .../MerchantTransactionBucketOperator.java         |   2 +-
 .../frauddetect/SlidingWindowSumKeyVal.java        |   6 +-
 .../frauddetect/SlidingWindowSumObject.java        |   3 +-
 .../frauddetect/TransactionStatsAggregator.java    |   4 +-
 .../operator/HdfsStringOutputOperator.java         |   3 +-
 examples/ftp/pom.xml                               |   2 +-
 .../org/apache/apex/examples/ftp/Application.java  |   2 +-
 examples/highlevelapi/pom.xml                      |   2 +-
 .../malhar/stream/sample/MinimalWordCount.java     |   2 +-
 .../malhar/stream/sample/WindowedWordCount.java    |   2 +-
 .../stream/sample/complete/AutoComplete.java       |   2 +-
 .../sample/complete/StreamingWordExtract.java      |   6 +-
 .../sample/complete/TopWikipediaSessions.java      |   2 +-
 .../stream/sample/complete/TrafficRoutes.java      |   2 +-
 .../sample/complete/TwitterAutoComplete.java       |   4 +-
 .../sample/cookbook/CombinePerKeyExamples.java     |   2 +-
 .../stream/sample/cookbook/MaxPerKeyExamples.java  |  14 +-
 .../stream/sample/cookbook/TriggerExample.java     |   3 +-
 .../stream/sample/WindowedWordCountTest.java       |   2 +-
 .../sample/complete/StreamingWordExtractTest.java  |   2 +-
 .../stream/sample/complete/TrafficRoutesTest.java  |   2 +-
 .../sample/cookbook/MaxPerKeyExamplesTest.java     |   2 +-
 examples/iteration/pom.xml                         |   2 +-
 .../apex/examples/iteration/Application.java       |   2 +-
 examples/jdbc/pom.xml                              |   2 +-
 .../FileToJdbcApp/FileToJdbcCsvParser.java         |  10 +-
 .../FileToJdbcApp/FileToJdbcCustomParser.java      |   6 +-
 .../JdbcIngest/FileLineOutputOperator.java         |   2 +-
 .../apex/examples/JdbcIngest/JdbcHDFSApp.java      |   8 +-
 .../examples/JdbcIngest/JdbcPollerApplication.java |   8 +-
 .../apex/examples/JdbcToJdbc/JdbcToJdbcApp.java    |  22 +-
 .../examples/FileToJdbcApp/ApplicationTest.java    |   2 +-
 .../apex/examples/JdbcToJdbc/JdbcOperatorTest.java |   6 +-
 examples/kafka/pom.xml                             |   2 +-
 .../kafka/exactlyonceoutput/ValidationToFile.java  |   3 +-
 .../examples/kafka/hdfs2kafka/Application.java     |   2 +-
 .../kafka/kafka2hdfs/LineOutputOperator.java       |   2 +-
 examples/machinedata/pom.xml                       |   2 +-
 .../apex/examples/machinedata/Application.java     |   4 +-
 .../machinedata/operator/CalculatorOperator.java   |   5 +-
 .../operator/MachineInfoAveragingOperator.java     |   5 +-
 .../MachineInfoAveragingPrerequisitesOperator.java |   2 +-
 .../operator/MachineInfoAveragingUnifier.java      |   3 +-
 .../machinedata/CalculatorOperatorTest.java        |   8 +-
 examples/mobile/pom.xml                            |   2 +-
 .../apache/apex/examples/mobile/Application.java   |  10 +-
 .../examples/mobile/PhoneMovementGenerator.java    |   4 +-
 .../apex/examples/mobile/ApplicationTest.java      |  10 +-
 examples/mrmonitor/pom.xml                         |   2 +-
 .../apex/examples/mrmonitor/Application.java       |   4 +-
 .../mrmonitor/MRMonitoringApplication.java         |   5 +-
 .../mrmonitor/MrMonitoringApplicationTest.java     |   2 +-
 examples/mroperator/pom.xml                        |   2 +-
 .../mroperator/HdfsKeyValOutputOperator.java       |   4 +-
 .../apex/examples/mroperator/MapOperator.java      |   3 +-
 .../examples/mroperator/OutputCollectorImpl.java   |   2 +-
 .../apex/examples/mroperator/ReduceOperator.java   |   2 +-
 .../apex/examples/mroperator/MapOperatorTest.java  |   3 +-
 .../examples/mroperator/ReduceOperatorTest.java    |   5 +-
 examples/nyctaxi/pom.xml                           |   2 +-
 .../apache/apex/examples/nyctaxi/Application.java  |   6 +-
 .../apex/examples/nyctaxi/NycTaxiDataServer.java   |   2 +-
 .../examples/nyctaxi/NycTaxiZipFareExtractor.java  |   2 +-
 examples/partition/pom.xml                         |   2 +-
 .../org/apache/apex/examples/partition/Codec3.java |   2 +-
 examples/pi/pom.xml                                |   2 +-
 .../org/apache/apex/examples/pi/Application.java   |   6 +-
 .../apex/examples/pi/ApplicationAppData.java       |  14 +-
 .../apex/examples/pi/ApplicationWithScript.java    |  14 +-
 .../org/apache/apex/examples/pi/Calculator.java    |  22 +-
 .../pi/src/main/resources/app/PiJsonExample.json   |   4 +-
 examples/pom.xml                                   |   2 +-
 examples/r/pom.xml                                 |   2 +-
 .../examples/r/oldfaithful/FaithfulRScript.java    |   3 +-
 .../r/oldfaithful/OldFaithfulApplication.java      |   2 +-
 examples/recordReader/pom.xml                      |   2 +-
 .../apex/examples/recordReader/Application.java    |   4 +-
 examples/s3/pom.xml                                |   2 +-
 .../apache/apex/examples/s3Output/Application.java |   2 +-
 .../s3ToHdfsSync/S3ToHDFSSyncApplication.java      |   4 +-
 .../META-INF/properties-s3TupleOutput.xml          |   2 +-
 examples/sql/pom.xml                               |   2 +-
 .../sql/sample/FusionStyleSQLApplication.java      |   2 +-
 examples/throttle/pom.xml                          |   2 +-
 examples/transform/pom.xml                         |   2 +-
 .../transform/DynamicTransformApplication.java     |   4 +-
 .../transform/SimpleTransformApplication.java      |   2 +-
 examples/twitter/pom.xml                           |   2 +-
 .../twitter/KinesisHashtagsApplication.java        |  24 +-
 .../examples/twitter/TwitterDumpApplication.java   |   5 +-
 .../twitter/TwitterDumpHBaseApplication.java       |   5 +-
 .../twitter/TwitterTopCounterApplication.java      |  19 +-
 .../twitter/TwitterTopWordsApplication.java        |   4 +-
 .../TwitterTrendingHashtagsApplication.java        |   8 +-
 .../examples/twitter/TwitterTopCounterTest.java    |   3 +-
 .../apex/examples/twitter/TwitterTopWordsTest.java |   3 +-
 examples/uniquecount/pom.xml                       |   2 +-
 .../apex/examples/uniquecount/Application.java     |  12 +-
 .../apex/examples/uniquecount/CountVerifier.java   |   3 +-
 .../examples/uniquecount/RandomDataGenerator.java  |   3 +-
 .../apex/examples/uniquecount/RandomKeyValues.java |   3 +-
 .../examples/uniquecount/RandomKeysGenerator.java  |   2 +-
 .../uniquecount/UniqueKeyValCountExample.java      |   9 +-
 examples/wordcount/pom.xml                         |   2 +-
 .../apex/examples/wordcount/Application.java       |   8 +-
 .../wordcount/ApplicationWithQuerySupport.java     |  11 +-
 .../apache/apex/examples/wordcount/LineReader.java |   2 +-
 .../examples/wordcount/WordCountInputOperator.java |   2 +-
 .../apex/examples/wordcount/WordCountWriter.java   |   2 +-
 examples/yahoofinance/pom.xml                      |   2 +-
 .../yahoofinance/ApplicationWithDerbySQL.java      |   2 +-
 .../apex/examples/yahoofinance/StockTickInput.java |   2 +-
 .../yahoofinance/YahooFinanceApplication.java      |  40 +--
 .../yahoofinance/YahooFinanceCSVInputOperator.java |   2 +-
 .../src/test/resources/alert_create.json           |   6 +-
 flume/pom.xml                                      |   2 +-
 .../apex/malhar/flume/storage/HDFSStorage.java     |   3 +-
 hive/pom.xml                                       |   2 +-
 .../hive/AbstractFSRollingOutputOperator.java      |   4 +-
 .../apex/malhar/hive/FSPojoToHiveOperator.java     |  22 +-
 .../org/apache/apex/malhar/hive/HiveOperator.java  |   8 +-
 .../apache/apex/malhar/hive/HiveOutputModule.java  |   5 +-
 .../org/apache/apex/malhar/hive/HiveStore.java     |   4 +-
 .../apache/apex/malhar/hive/HiveStreamCodec.java   |   6 +-
 .../org/apache/apex/malhar/hive/package-info.java  |   2 +-
 .../apex/malhar/hive/FSRollingMapTestImpl.java     |   2 +-
 .../apache/apex/malhar/hive/FSRollingTestImpl.java |   2 +-
 .../org/apache/apex/malhar/hive/HiveMockTest.java  |  10 +-
 .../org/apache/apex/malhar/hive/HiveStoreTest.java |   2 +-
 kafka/kafka-common/pom.xml                         |   2 +-
 .../malhar/kafka/AbstractKafkaPartitioner.java     |   2 +-
 kafka/kafka010/pom.xml                             |   2 +-
 .../apex/malhar/kafka/KafkaOperatorTestBase.java   |   2 +-
 kafka/kafka09/pom.xml                              |   2 +-
 .../apex/malhar/kafka/KafkaOperatorTestBase.java   |   2 +-
 .../apex/malhar/kafka/KafkaOutputOperatorTest.java |   2 +-
 kafka/pom.xml                                      |   2 +-
 kudu/pom.xml                                       |   2 +-
 .../malhar/kudu/AbstractKuduInputOperator.java     |   2 +-
 .../malhar/kudu/AbstractKuduOutputOperator.java    |   2 +-
 .../apex/malhar/kudu/ApexKuduConnection.java       |   7 +-
 .../partitioner/AbstractKuduInputPartitioner.java  |   2 +-
 .../kudu/scanner/KuduPartitionScannerCallable.java |   3 +-
 .../kudu/IncrementalStepScanInputOperatorTest.java |   4 +-
 .../apex/malhar/kudu/KuduClientTestCommons.java    |   3 +-
 .../KuduCreateUpdateDeleteOutputOperatorTest.java  |   4 +-
 .../apex/malhar/kudu/KuduInputOperatorCommons.java |   4 +-
 .../partitioner/KuduOneToManyPartitionerTest.java  |   2 +-
 .../partitioner/KuduOneToOnePartitionerTest.java   |   2 +-
 library/pom.xml                                    |   2 +-
 .../apache/apex/malhar/lib/algo/BottomNMap.java    |   6 +-
 .../apex/malhar/lib/algo/BottomNUnifier.java       |   5 +-
 .../apex/malhar/lib/algo/BottomNUniqueMap.java     |   6 +-
 .../malhar/lib/algo/CompareExceptCountMap.java     |   6 +-
 .../org/apache/apex/malhar/lib/algo/Distinct.java  |   6 +-
 .../apache/apex/malhar/lib/algo/FilterValues.java  |   2 +-
 .../apache/apex/malhar/lib/algo/InsertSort.java    |   6 +-
 .../lib/algo/LeastFrequentKeyArrayUnifier.java     |   2 +-
 .../malhar/lib/algo/LeastFrequentKeyUnifier.java   |   2 +-
 .../apex/malhar/lib/algo/LeastFrequentValue.java   |   6 +-
 .../apache/apex/malhar/lib/algo/MatchAllMap.java   |   8 +-
 .../apache/apex/malhar/lib/algo/MatchAnyMap.java   |   8 +-
 .../org/apache/apex/malhar/lib/algo/MatchMap.java  |   8 +-
 .../org/apache/apex/malhar/lib/algo/MergeSort.java |   2 +-
 .../apex/malhar/lib/algo/MergeSortNumber.java      |   2 +-
 .../apex/malhar/lib/algo/MostFrequentValue.java    |  10 +-
 .../java/org/apache/apex/malhar/lib/algo/TopN.java |   6 +-
 .../apache/apex/malhar/lib/algo/TopNUnique.java    |   6 +-
 .../apache/apex/malhar/lib/algo/UniqueCounter.java |   7 +-
 .../apex/malhar/lib/algo/UniqueValueCount.java     |   5 +-
 .../apache/apex/malhar/lib/algo/package-info.java  |   2 +-
 .../malhar/lib/appdata/AbstractAppDataServer.java  |   1 -
 .../apache/apex/malhar/lib/appdata/QueueUtils.java |   2 +-
 .../apache/apex/malhar/lib/appdata/StoreUtils.java |   5 +-
 .../datastructs/CacheLRUSynchronousFlush.java      |   2 +-
 .../lib/appdata/datastructs/DimensionalTable.java  |   2 +-
 .../appdata/datastructs/NOPCacheFlushListener.java |   4 +-
 .../malhar/lib/appdata/gpo/GPOByteArrayList.java   |   2 +-
 .../apex/malhar/lib/appdata/gpo/GPOGetters.java    |  20 +-
 .../apex/malhar/lib/appdata/gpo/GPOMutable.java    |  10 +-
 .../apex/malhar/lib/appdata/gpo/GPOType.java       |  61 +++--
 .../apex/malhar/lib/appdata/gpo/GPOUtils.java      |  31 ++-
 .../apache/apex/malhar/lib/appdata/gpo/Serde.java  |   2 +-
 .../lib/appdata/gpo/SerdeFieldsDescriptor.java     |   7 +-
 .../lib/appdata/gpo/SerdeListGPOMutable.java       |   5 +-
 .../malhar/lib/appdata/gpo/SerdeListPrimitive.java |   5 +-
 .../malhar/lib/appdata/gpo/SerdeListString.java    |   2 +-
 .../malhar/lib/appdata/gpo/SerdeMapPrimitive.java  |   5 +-
 .../lib/appdata/gpo/SerdeObjectPayloadFix.java     |   2 +-
 .../apex/malhar/lib/appdata/package-info.java      |   2 +-
 .../query/AbstractWindowEndQueueManager.java       |   7 +-
 .../query/AppDataWindowEndQueueManager.java        |   7 +-
 .../apex/malhar/lib/appdata/query/QueryBundle.java |   2 +-
 .../malhar/lib/appdata/query/QueryExecutor.java    |   2 +-
 .../appdata/query/QueryManagerAsynchronous.java    |   7 +-
 .../lib/appdata/query/QueryManagerSynchronous.java |   2 +-
 .../apex/malhar/lib/appdata/query/QueueList.java   |   2 +-
 .../malhar/lib/appdata/query/QueueManager.java     |   2 +-
 .../lib/appdata/query/SimpleDoneQueueManager.java  |   5 +-
 .../lib/appdata/query/SimpleQueueManager.java      |   5 +-
 .../lib/appdata/query/WindowBoundedService.java    |   2 +-
 .../lib/appdata/query/WindowEndQueueManager.java   |   5 +-
 .../query/serde/CustomMessageDeserializer.java     |   4 +-
 .../query/serde/CustomMessageSerializer.java       |   6 +-
 .../query/serde/CustomMessageValidator.java        |   4 +-
 .../query/serde/DataQuerySnapshotDeserializer.java |  12 +-
 .../query/serde/DataQuerySnapshotValidator.java    |  12 +-
 .../query/serde/DataResultSnapshotSerializer.java  |  16 +-
 .../query/serde/MessageDeserializerFactory.java    |   6 +-
 .../query/serde/MessageDeserializerInfo.java       |   2 +-
 .../query/serde/MessageSerializerFactory.java      |   8 +-
 .../appdata/query/serde/MessageSerializerInfo.java |   2 +-
 .../lib/appdata/query/serde/MessageType.java       |   2 +-
 .../appdata/query/serde/MessageValidatorInfo.java  |   2 +-
 .../query/serde/SchemaQueryDeserializer.java       |  10 +-
 .../query/serde/SimpleDataDeserializer.java        |   4 +-
 .../appdata/query/serde/SimpleDataSerializer.java  |   6 +-
 .../appdata/query/serde/SimpleDataValidator.java   |   4 +-
 .../lib/appdata/schemas/CustomTimeBucket.java      |   2 +-
 .../lib/appdata/schemas/DataQuerySnapshot.java     |  14 +-
 .../lib/appdata/schemas/DataResultSnapshot.java    |  12 +-
 .../schemas/DimensionalConfigurationSchema.java    |   4 +-
 .../lib/appdata/schemas/DimensionalSchema.java     |   2 +-
 .../apex/malhar/lib/appdata/schemas/Fields.java    |   2 +-
 .../lib/appdata/schemas/FieldsDescriptor.java      |   8 +-
 .../apex/malhar/lib/appdata/schemas/JSONType.java  |   2 +-
 .../apex/malhar/lib/appdata/schemas/Message.java   |   2 +-
 .../apex/malhar/lib/appdata/schemas/QRBase.java    |   2 +-
 .../apex/malhar/lib/appdata/schemas/Query.java     |   2 +-
 .../apex/malhar/lib/appdata/schemas/Result.java    |   2 +-
 .../lib/appdata/schemas/ResultFormatter.java       |   2 +-
 .../apex/malhar/lib/appdata/schemas/Schema.java    |   2 +-
 .../malhar/lib/appdata/schemas/SchemaQuery.java    |  12 +-
 .../malhar/lib/appdata/schemas/SchemaRegistry.java |   2 +-
 .../appdata/schemas/SchemaRegistryMultiple.java    |   8 +-
 .../lib/appdata/schemas/SchemaRegistrySingle.java  |   2 +-
 .../malhar/lib/appdata/schemas/SchemaResult.java   |   8 +-
 .../appdata/schemas/SchemaResultSerializer.java    |   4 +-
 .../malhar/lib/appdata/schemas/SchemaUtils.java    |   2 +-
 .../malhar/lib/appdata/schemas/SnapshotSchema.java |   2 +-
 .../malhar/lib/appdata/schemas/TimeBucket.java     |   2 +-
 .../apex/malhar/lib/appdata/schemas/Type.java      |   2 +-
 .../snapshot/AbstractAppDataSnapshotServer.java    |  36 +--
 .../appdata/snapshot/AppDataSnapshotServerMap.java |   8 +-
 .../snapshot/AppDataSnapshotServerPOJO.java        |   8 +-
 .../lib/bandwidth/BandwidthLimitingOperator.java   |   2 +-
 .../malhar/lib/bandwidth/BandwidthManager.java     |   2 +-
 .../malhar/lib/bandwidth/BandwidthPartitioner.java |   2 +-
 .../lib/codec/JavaSerializationStreamCodec.java    |   2 +-
 .../apex/malhar/lib/codec/KryoJdkContainer.java    |   2 +-
 .../lib/codec/KryoSerializableStreamCodec.java     |   2 +-
 .../apache/apex/malhar/lib/codec/package-info.java |   2 +-
 .../ByteArrayToStringConverterOperator.java        |   2 +-
 .../apex/malhar/lib/converter/Converter.java       |   2 +-
 .../converter/MapToKeyHashValuePairConverter.java  |   5 +-
 .../lib/converter/MapToKeyValuePairConverter.java  |   5 +-
 .../StringValueToNumberConverterForMap.java        |   2 +-
 .../apex/malhar/lib/converter/package-info.java    |   2 +-
 .../apex/malhar/lib/counters/BasicCounters.java    |   2 +-
 .../apex/malhar/lib/counters/package-info.java     |   2 +-
 ...TransactionableKeyValueStoreOutputOperator.java |   2 +-
 ...ggregateTransactionableStoreOutputOperator.java |   2 +-
 ...actBatchTransactionableStoreOutputOperator.java |   2 +-
 .../lib/db/AbstractKeyValueStoreInputOperator.java |   2 +-
 ...TransactionableKeyValueStoreOutputOperator.java |   2 +-
 ...PassThruTransactionableStoreOutputOperator.java |   2 +-
 .../malhar/lib/db/AbstractStoreInputOperator.java  |   2 +-
 .../malhar/lib/db/AbstractStoreOutputOperator.java |   2 +-
 ...AbstractTransactionableStoreOutputOperator.java |   2 +-
 .../org/apache/apex/malhar/lib/db/Connectable.java |   2 +-
 .../apache/apex/malhar/lib/db/KeyValueStore.java   |   2 +-
 .../apache/apex/malhar/lib/db/Transactionable.java |   2 +-
 .../lib/db/TransactionableKeyValueStore.java       |   2 +-
 .../apex/malhar/lib/db/TransactionableStore.java   |   2 +-
 .../cache/AbstractDBLookupCacheBackedOperator.java |   7 +-
 .../apex/malhar/lib/db/cache/CacheManager.java     |   5 +-
 .../apex/malhar/lib/db/cache/CacheStore.java       |   2 +-
 .../apex/malhar/lib/db/cache/package-info.java     |   2 +-
 .../lib/db/jdbc/AbstractJdbcInputOperator.java     |   6 +-
 ...tJdbcNonTransactionableBatchOutputOperator.java |   2 +-
 ...stractJdbcNonTransactionableOutputOperator.java |   5 +-
 .../db/jdbc/AbstractJdbcPOJOOutputOperator.java    |  21 +-
 .../lib/db/jdbc/AbstractJdbcPollInputOperator.java |   8 +-
 .../AbstractJdbcTransactionableOutputOperator.java |   5 +-
 .../lib/db/jdbc/JDBCDimensionalOutputOperator.java |  12 +-
 .../lib/db/jdbc/JDBCLookupCacheBackedOperator.java |   6 +-
 .../apex/malhar/lib/db/jdbc/JdbcFieldInfo.java     |   4 +-
 .../lib/db/jdbc/JdbcNonTransactionalStore.java     |   2 +-
 .../malhar/lib/db/jdbc/JdbcPOJOInputOperator.java  |   7 +-
 .../lib/db/jdbc/JdbcPOJOInsertOutputOperator.java  |   5 +-
 .../db/jdbc/JdbcPOJONonInsertOutputOperator.java   |   2 +-
 .../lib/db/jdbc/JdbcPOJOPollInputOperator.java     |   8 +-
 .../malhar/lib/db/jdbc/JdbcPollInputOperator.java  |   2 +-
 .../apache/apex/malhar/lib/db/jdbc/JdbcStore.java  |   5 +-
 .../malhar/lib/db/jdbc/JdbcTransactionalStore.java |   4 +-
 .../apex/malhar/lib/db/jdbc/package-info.java      |   2 +-
 .../apache/apex/malhar/lib/db/package-info.java    |   2 +-
 .../RedshiftJdbcTransactionableOutputOperator.java |   2 +-
 .../lib/db/redshift/RedshiftOutputModule.java      |   2 +-
 .../apex/malhar/lib/dedup/AbstractDeduper.java     |   2 +-
 .../malhar/lib/dedup/BoundedDedupOperator.java     |   5 +-
 .../apex/malhar/lib/dedup/DeduperStreamCodec.java  |   6 +-
 .../malhar/lib/dedup/TimeBasedDedupOperator.java   |   4 +-
 .../lib/dimensions/CustomTimeBucketRegistry.java   |   4 +-
 .../dimensions/DimensionsConversionContext.java    |   4 +-
 .../lib/dimensions/DimensionsDescriptor.java       |  14 +-
 .../malhar/lib/dimensions/DimensionsEvent.java     |   5 +-
 .../aggregator/AbstractCompositeAggregator.java    |   3 +-
 .../aggregator/AbstractIncrementalAggregator.java  |   9 +-
 .../aggregator/AbstractTopBottomAggregator.java    |   5 +-
 .../dimensions/aggregator/AggregatorAverage.java   |   9 +-
 .../lib/dimensions/aggregator/AggregatorCount.java |   8 +-
 .../dimensions/aggregator/AggregatorCumSum.java    |  16 +-
 .../lib/dimensions/aggregator/AggregatorFirst.java |   6 +-
 .../lib/dimensions/aggregator/AggregatorLast.java  |   6 +-
 .../lib/dimensions/aggregator/AggregatorMax.java   |   8 +-
 .../lib/dimensions/aggregator/AggregatorMin.java   |   8 +-
 .../dimensions/aggregator/AggregatorRegistry.java  |   2 +-
 .../lib/dimensions/aggregator/AggregatorSum.java   |   8 +-
 .../lib/dimensions/aggregator/AggregatorUtils.java |  12 +-
 .../dimensions/aggregator/CompositeAggregator.java |   5 +-
 .../aggregator/IncrementalAggregator.java          |   7 +-
 .../lib/dimensions/aggregator/OTFAggregator.java   |   4 +-
 .../apex/malhar/lib/expression/Expression.java     |   2 +-
 .../lib/expression/JavaExpressionParser.java       |   4 +-
 .../apex/malhar/lib/fileaccess/DTFileReader.java   |   2 +-
 .../apex/malhar/lib/fileaccess/FileAccess.java     |   2 +-
 .../malhar/lib/fileaccess/FileAccessFSImpl.java    |   2 +-
 .../apex/malhar/lib/fileaccess/TFileImpl.java      |   2 +-
 .../apex/malhar/lib/fileaccess/TFileReader.java    |   2 +-
 .../apex/malhar/lib/fileaccess/TFileWriter.java    |   2 +-
 .../apex/malhar/lib/fileaccess/package-info.java   |   2 +-
 .../apex/malhar/lib/filter/FilterOperator.java     |   6 +-
 .../apex/malhar/lib/formatter/Formatter.java       |   5 +-
 .../apex/malhar/lib/formatter/JsonFormatter.java   |   2 +-
 .../apex/malhar/lib/formatter/XmlFormatter.java    |   2 +-
 .../apache/apex/malhar/lib/fs/FSRecordReader.java  |   6 +-
 .../apex/malhar/lib/fs/FSRecordReaderModule.java   |   4 +-
 .../malhar/lib/fs/GenericFileOutputOperator.java   |   5 +-
 .../malhar/lib/fs/LineByLineFileInputOperator.java |   3 +-
 .../malhar/lib/fs/s3/S3BlockUploadOperator.java    |   4 +-
 .../lib/fs/s3/S3InitiateFileUploadOperator.java    |   2 +-
 .../apex/malhar/lib/fs/s3/S3OutputModule.java      |   6 +-
 .../apache/apex/malhar/lib/fs/s3/S3Reconciler.java |   2 +-
 .../lib/fs/s3/S3ReconcilerQueuePartitioner.java    |   3 +-
 .../apex/malhar/lib/fs/s3/S3RecordReader.java      |   5 +-
 .../malhar/lib/fs/s3/S3RecordReaderModule.java     |   3 +-
 .../apex/malhar/lib/fs/s3/S3TupleOutputModule.java |   4 +-
 .../apache/apex/malhar/lib/function/Function.java  |   3 +-
 .../malhar/lib/io/AbstractFTPInputOperator.java    |   4 +-
 .../io/AbstractHttpGetMultiValuedMapOperator.java  |   2 +-
 .../malhar/lib/io/AbstractHttpGetOperator.java     |   2 +-
 .../malhar/lib/io/AbstractHttpInputOperator.java   |   2 +-
 .../apex/malhar/lib/io/AbstractHttpOperator.java   |   2 +-
 .../io/AbstractKeyValueStoreOutputOperator.java    |   5 +-
 .../malhar/lib/io/AbstractSocketInputOperator.java |   2 +-
 .../apex/malhar/lib/io/ApacheGenRandomLogs.java    |   2 +-
 .../io/CollectionMultiConsoleOutputOperator.java   |   2 +-
 .../apex/malhar/lib/io/ConsoleOutputOperator.java  |   2 +-
 .../apex/malhar/lib/io/HttpGetMapOperator.java     |   2 +-
 .../malhar/lib/io/HttpJsonChunksInputOperator.java |   2 +-
 .../apex/malhar/lib/io/HttpLinesInputOperator.java |   2 +-
 .../apex/malhar/lib/io/HttpOutputOperator.java     |   2 +-
 .../apex/malhar/lib/io/HttpPostOutputOperator.java |   2 +-
 .../malhar/lib/io/IdempotentStorageManager.java    |   4 +-
 .../lib/io/MapMultiConsoleOutputOperator.java      |   2 +-
 .../malhar/lib/io/PubSubWebSocketAppDataQuery.java |   7 +-
 .../lib/io/PubSubWebSocketAppDataResult.java       |   2 +-
 .../lib/io/PubSubWebSocketInputOperator.java       |   2 +-
 .../lib/io/PubSubWebSocketOutputOperator.java      |   2 +-
 .../lib/io/SimpleSinglePortInputOperator.java      |   2 +-
 .../apex/malhar/lib/io/SmtpOutputOperator.java     |   2 +-
 .../apex/malhar/lib/io/WebSocketInputOperator.java |   2 +-
 .../malhar/lib/io/WebSocketOutputOperator.java     |   2 +-
 .../lib/io/WebSocketServerInputOperator.java       |   2 +-
 .../apex/malhar/lib/io/WidgetOutputOperator.java   |   2 +-
 .../malhar/lib/io/block/AbstractBlockReader.java   |   6 +-
 .../malhar/lib/io/block/AbstractFSBlockReader.java |   2 +-
 .../apex/malhar/lib/io/block/BlockMetadata.java    |   2 +-
 .../apex/malhar/lib/io/block/BlockWriter.java      |   6 +-
 .../apex/malhar/lib/io/block/FSSliceReader.java    |   2 +-
 .../apex/malhar/lib/io/block/PartFileWriter.java   |   4 +-
 .../apex/malhar/lib/io/block/ReaderContext.java    |   2 +-
 .../apex/malhar/lib/io/block/package-info.java     |   2 +-
 .../lib/io/fs/AbstractFileInputOperator.java       |   7 +-
 .../lib/io/fs/AbstractFileOutputOperator.java      |   4 +-
 .../malhar/lib/io/fs/AbstractFileSplitter.java     |   5 +-
 .../apex/malhar/lib/io/fs/AbstractReconciler.java  |   2 +-
 .../io/fs/AbstractSingleFileOutputOperator.java    |   2 +-
 .../io/fs/AbstractThroughputFileInputOperator.java |   4 +-
 .../io/fs/AbstractWindowFileOutputOperator.java    |   2 +-
 .../apex/malhar/lib/io/fs/FSInputModule.java       |  13 +-
 .../apache/apex/malhar/lib/io/fs/FileMerger.java   |   4 +-
 .../apache/apex/malhar/lib/io/fs/FileSplitter.java |   6 +-
 .../apex/malhar/lib/io/fs/FileSplitterBase.java    |   2 +-
 .../apex/malhar/lib/io/fs/FileSplitterInput.java   |   2 +-
 .../apache/apex/malhar/lib/io/fs/FileStitcher.java |   8 +-
 .../apex/malhar/lib/io/fs/FilterStreamCodec.java   |   2 +-
 .../apex/malhar/lib/io/fs/FilterStreamContext.java |   2 +-
 .../malhar/lib/io/fs/FilterStreamProvider.java     |   2 +-
 .../apex/malhar/lib/io/fs/HDFSFileCopyModule.java  |  10 +-
 .../apex/malhar/lib/io/fs/HDFSFileMerger.java      |   4 +-
 .../apex/malhar/lib/io/fs/S3BlockReader.java       |   8 +-
 .../apex/malhar/lib/io/fs/S3InputModule.java       |   4 +-
 .../apache/apex/malhar/lib/io/fs/Synchronizer.java |  10 +-
 .../apex/malhar/lib/io/fs/TailFsInputOperator.java |   2 +-
 .../apache/apex/malhar/lib/io/fs/package-info.java |   2 +-
 .../lib/io/jms/AbstractJMSInputOperator.java       |   4 +-
 .../lib/io/jms/AbstractJMSOutputOperator.java      |   2 +-
 .../jms/AbstractJMSSinglePortOutputOperator.java   |   2 +-
 .../lib/io/jms/FSPsuedoTransactionableStore.java   |   2 +-
 .../org/apache/apex/malhar/lib/io/jms/JMSBase.java |   2 +-
 .../lib/io/jms/JMSBaseTransactionableStore.java    |   4 +-
 .../lib/io/jms/JMSMultiPortOutputOperator.java     |   2 +-
 .../malhar/lib/io/jms/JMSObjectInputOperator.java  |   2 +-
 .../malhar/lib/io/jms/JMSStringInputOperator.java  |   2 +-
 .../malhar/lib/io/jms/JMSTransactionableStore.java |   2 +-
 .../apex/malhar/lib/io/jms/package-info.java       |   2 +-
 .../apache/apex/malhar/lib/io/package-info.java    |   2 +-
 .../apex/malhar/lib/join/AbstractJoinOperator.java |   2 +-
 .../AbstractManagedStateInnerJoinOperator.java     |   3 +-
 .../apex/malhar/lib/join/AntiJoinOperator.java     |   6 +-
 .../org/apache/apex/malhar/lib/join/Bucket.java    |   2 +-
 .../apache/apex/malhar/lib/join/InMemoryStore.java |   2 +-
 .../org/apache/apex/malhar/lib/join/JoinStore.java |   2 +-
 .../apex/malhar/lib/join/JoinStreamCodec.java      |   4 +-
 .../apex/malhar/lib/join/MapJoinOperator.java      |   2 +-
 .../malhar/lib/join/POJOInnerJoinOperator.java     |   2 +-
 .../apex/malhar/lib/join/POJOJoinOperator.java     |   5 +-
 .../apex/malhar/lib/join/SemiJoinOperator.java     |   6 +-
 .../apex/malhar/lib/join/TimeBasedStore.java       |   2 +-
 .../org/apache/apex/malhar/lib/join/TimeEvent.java |   2 +-
 .../apache/apex/malhar/lib/join/TimeEventImpl.java |   2 +-
 .../lib/logs/ApacheLogParseMapOutputOperator.java  |   2 +-
 .../malhar/lib/logs/ApacheLogParseOperator.java    |   2 +-
 .../lib/logs/ApacheVirtualLogParseOperator.java    |   4 +-
 .../lib/logs/DimensionAggregationUnifier.java      |   2 +-
 .../apex/malhar/lib/logs/DimensionObject.java      |   2 +-
 .../lib/logs/FilteredLineToTokenArrayList.java     |   2 +-
 .../lib/logs/FilteredLineToTokenHashMap.java       |   2 +-
 .../lib/logs/FilteredLineTokenizerKeyVal.java      |   2 +-
 .../apex/malhar/lib/logs/InformationExtractor.java |   2 +-
 .../apex/malhar/lib/logs/LineToTokenArrayList.java |   8 +-
 .../apex/malhar/lib/logs/LineToTokenHashMap.java   |   8 +-
 .../apache/apex/malhar/lib/logs/LineTokenizer.java |   5 +-
 .../apex/malhar/lib/logs/LineTokenizerKeyVal.java  |   7 +-
 .../lib/logs/MultiWindowDimensionAggregation.java  |   4 +-
 .../malhar/lib/logs/RegexMatchMapOperator.java     |   2 +-
 .../apache/apex/malhar/lib/logs/package-info.java  |   2 +-
 .../malhar/lib/math/AbstractAggregateCalc.java     |   2 +-
 .../apex/malhar/lib/math/AbstractOutput.java       |   2 +-
 .../lib/math/AbstractXmlCartesianProduct.java      |   5 +-
 .../math/AbstractXmlKeyValueCartesianProduct.java  |   2 +-
 .../org/apache/apex/malhar/lib/math/Average.java   |   5 +-
 .../apache/apex/malhar/lib/math/AverageKeyVal.java |   6 +-
 .../org/apache/apex/malhar/lib/math/Division.java  |   2 +-
 .../apex/malhar/lib/math/LogicalCompare.java       |   2 +-
 .../malhar/lib/math/LogicalCompareToConstant.java  |   2 +-
 .../org/apache/apex/malhar/lib/math/Margin.java    |   5 +-
 .../apache/apex/malhar/lib/math/MarginKeyVal.java  |   6 +-
 .../org/apache/apex/malhar/lib/math/MarginMap.java |   6 +-
 .../java/org/apache/apex/malhar/lib/math/Max.java  |   5 +-
 .../org/apache/apex/malhar/lib/math/MaxKeyVal.java |   7 +-
 .../java/org/apache/apex/malhar/lib/math/Min.java  |   5 +-
 .../org/apache/apex/malhar/lib/math/MinKeyVal.java |   7 +-
 .../apex/malhar/lib/math/MultiplyByConstant.java   |   2 +-
 .../org/apache/apex/malhar/lib/math/Range.java     |   9 +-
 .../apache/apex/malhar/lib/math/RangeKeyVal.java   |  11 +-
 .../apex/malhar/lib/math/RunningAverage.java       |   2 +-
 .../org/apache/apex/malhar/lib/math/Sigma.java     |   2 +-
 .../lib/math/SingleVariableAbstractCalculus.java   |   2 +-
 .../apex/malhar/lib/math/SquareCalculus.java       |   2 +-
 .../java/org/apache/apex/malhar/lib/math/Sum.java  |   7 +-
 .../org/apache/apex/malhar/lib/math/SumKeyVal.java |   6 +-
 .../math/XmlKeyValueStringCartesianProduct.java    |   2 +-
 .../apache/apex/malhar/lib/math/package-info.java  |   2 +-
 .../apex/malhar/lib/metric/AvgAggregator.java      |   2 +-
 .../malhar/lib/metric/max/DoubleMaxAggregator.java |   2 +-
 .../malhar/lib/metric/max/FloatMaxAggregator.java  |   2 +-
 .../malhar/lib/metric/max/IntMaxAggregator.java    |   2 +-
 .../malhar/lib/metric/max/LongMaxAggregator.java   |   2 +-
 .../malhar/lib/metric/min/DoubleMinAggregator.java |   2 +-
 .../malhar/lib/metric/min/FloatMinAggregator.java  |   2 +-
 .../malhar/lib/metric/min/IntMinAggregator.java    |   2 +-
 .../malhar/lib/metric/min/LongMinAggregator.java   |   2 +-
 .../apex/malhar/lib/metric/package-info.java       |   2 +-
 .../lib/multiwindow/AbstractSlidingWindow.java     |   2 +-
 .../multiwindow/AbstractSlidingWindowKeyVal.java   |   7 +-
 .../lib/multiwindow/MultiWindowRangeKeyVal.java    |   8 +-
 .../lib/multiwindow/MultiWindowSumKeyVal.java      |   7 +-
 .../lib/multiwindow/SimpleMovingAverage.java       |   5 +-
 .../lib/multiwindow/SimpleMovingAverageObject.java |   2 +-
 .../malhar/lib/multiwindow/SortedMovingWindow.java |   2 +-
 .../apex/malhar/lib/multiwindow/package-info.java  |   2 +-
 .../org/apache/apex/malhar/lib/parser/Parser.java  |   4 +-
 .../apache/apex/malhar/lib/parser/XmlParser.java   |   5 +-
 .../apex/malhar/lib/parser/package-info.java       |   2 +-
 .../StatelessThroughputBasedPartitioner.java       |   2 +-
 .../StatsAwareStatelessPartitioner.java            |   2 +-
 .../apex/malhar/lib/partitioner/package-info.java  |   2 +-
 .../malhar/lib/projection/ProjectionOperator.java  |   5 +-
 .../apex/malhar/lib/script/JavaScriptOperator.java |   2 +-
 .../apex/malhar/lib/script/ScriptOperator.java     |   2 +-
 .../apex/malhar/lib/script/package-info.java       |   2 +-
 .../state/managed/AbstractManagedStateImpl.java    |   7 +-
 .../apex/malhar/lib/state/managed/Bucket.java      |   2 +-
 .../lib/state/managed/BucketsFileSystem.java       |   2 +-
 .../lib/state/managed/ManagedStateContext.java     |   3 +-
 .../state/managed/ManagedTimeStateMultiValue.java  |   2 +-
 .../state/managed/ManagedTimeUnifiedStateImpl.java |   2 +-
 .../apex/malhar/lib/statistics/MedianOperator.java |   2 +-
 .../apex/malhar/lib/statistics/ModeOperator.java   |   2 +-
 .../malhar/lib/statistics/StandardDeviation.java   |   2 +-
 .../lib/statistics/WeightedMeanOperator.java       |   6 +-
 .../apex/malhar/lib/statistics/package-info.java   |   2 +-
 .../apex/malhar/lib/stream/AbstractAggregator.java |   2 +-
 .../malhar/lib/stream/ArrayListAggregator.java     |   2 +-
 .../apex/malhar/lib/stream/ArrayListToItem.java    |   5 +-
 .../apex/malhar/lib/stream/ConsolidatorKeyVal.java |   5 +-
 .../org/apache/apex/malhar/lib/stream/Counter.java |   2 +-
 .../org/apache/apex/malhar/lib/stream/DevNull.java |   2 +-
 .../apex/malhar/lib/stream/DevNullCounter.java     |   2 +-
 .../malhar/lib/stream/HashMapToKeyValPair.java     |   7 +-
 .../malhar/lib/stream/JsonByteArrayOperator.java   |   2 +-
 .../malhar/lib/stream/KeyValPairToHashMap.java     |   7 +-
 .../apex/malhar/lib/stream/RoundRobinHashMap.java  |   5 +-
 .../apex/malhar/lib/stream/StreamDuplicater.java   |   5 +-
 .../apex/malhar/lib/stream/StreamMerger.java       |   2 +-
 .../apex/malhar/lib/stream/package-info.java       |   2 +-
 .../lib/streamquery/condition/Condition.java       |   2 +-
 .../condition/JoinColumnEqualCondition.java        |   2 +-
 .../malhar/lib/streamquery/index/ColumnIndex.java  |   2 +-
 .../apex/malhar/lib/streamquery/index/Index.java   |   2 +-
 .../malhar/lib/testbench/ArrayListTestSink.java    |   2 +-
 .../malhar/lib/testbench/CollectorTestSink.java    |   2 +-
 .../malhar/lib/testbench/CompareFilterTuples.java  |   2 +-
 .../lib/testbench/CountAndLastTupleTestSink.java   |   2 +-
 .../apex/malhar/lib/testbench/CountOccurance.java  |   2 +-
 .../apex/malhar/lib/testbench/CountTestSink.java   |   2 +-
 .../apex/malhar/lib/testbench/EventClassifier.java |   2 +-
 .../EventClassifierNumberToHashDouble.java         |   2 +-
 .../apex/malhar/lib/testbench/EventGenerator.java  |   2 +-
 .../malhar/lib/testbench/EventIncrementer.java     |   5 +-
 .../malhar/lib/testbench/FilterClassifier.java     |   2 +-
 .../lib/testbench/FilteredEventClassifier.java     |   2 +-
 .../apex/malhar/lib/testbench/HashTestSink.java    |   2 +-
 .../malhar/lib/testbench/HttpStatusFilter.java     |   2 +-
 .../apex/malhar/lib/testbench/KeyValSum.java       |   2 +-
 .../malhar/lib/testbench/RandomEventGenerator.java |   2 +-
 .../malhar/lib/testbench/RandomWordGenerator.java  |   2 +-
 .../apex/malhar/lib/testbench/RedisSumOper.java    |   2 +-
 .../malhar/lib/testbench/SeedEventClassifier.java  |   2 +-
 .../malhar/lib/testbench/SeedEventGenerator.java   |   5 +-
 .../apex/malhar/lib/testbench/SumTestSink.java     |   2 +-
 .../malhar/lib/testbench/ThroughputCounter.java    |   2 +-
 .../apex/malhar/lib/testbench/TopOccurrence.java   |   2 +-
 .../apex/malhar/lib/testbench/package-info.java    |   2 +-
 .../malhar/lib/transform/TransformOperator.java    |   6 +-
 .../malhar/lib/util/AbstractBaseFrequentKey.java   |   2 +-
 .../lib/util/AbstractBaseFrequentKeyValueMap.java  |   2 +-
 .../malhar/lib/util/AbstractBaseMatchOperator.java |   2 +-
 .../util/AbstractBaseNNonUniqueOperatorMap.java    |   2 +-
 .../malhar/lib/util/AbstractBaseNOperatorMap.java  |   2 +-
 .../lib/util/AbstractBaseNUniqueOperatorMap.java   |   2 +-
 .../malhar/lib/util/AbstractBaseSortOperator.java  |   2 +-
 .../util/AbstractDimensionTimeBucketOperator.java  |   2 +-
 .../lib/util/AbstractKeyValueStorageAgent.java     |   2 +-
 .../malhar/lib/util/ActiveMQMessageListener.java   |   2 +-
 .../lib/util/ActiveMQMultiTypeMessageListener.java |   2 +-
 .../apex/malhar/lib/util/ArrayHashMapFrequent.java |   2 +-
 .../lib/util/BaseFilteredKeyValueOperator.java     |   2 +-
 .../apex/malhar/lib/util/BaseKeyOperator.java      |   2 +-
 .../apex/malhar/lib/util/BaseKeyValueOperator.java |   5 +-
 .../apex/malhar/lib/util/BaseLineTokenizer.java    |   2 +-
 .../apex/malhar/lib/util/BaseMatchOperator.java    |   2 +-
 .../lib/util/BaseNumberKeyValueOperator.java       |   2 +-
 .../malhar/lib/util/BaseNumberValueOperator.java   |   2 +-
 .../apex/malhar/lib/util/BaseUniqueKeyCounter.java |   2 +-
 .../malhar/lib/util/BaseUniqueKeyValueCounter.java |   2 +-
 .../lib/util/DimensionTimeBucketSumOperator.java   |   2 +-
 .../org/apache/apex/malhar/lib/util/FieldInfo.java |   2 +-
 .../apex/malhar/lib/util/FieldValueGenerator.java  |   6 +-
 .../apex/malhar/lib/util/FilterOperator.java       |   2 +-
 .../org/apache/apex/malhar/lib/util/HighLow.java   |   2 +-
 .../malhar/lib/util/JavaScriptFilterOperator.java  |   2 +-
 .../apex/malhar/lib/util/KeyHashValPair.java       |   2 +-
 .../apache/apex/malhar/lib/util/KeyValPair.java    |   2 +-
 .../apex/malhar/lib/util/KryoCloneUtils.java       |   2 +-
 .../org/apache/apex/malhar/lib/util/PojoUtils.java |   9 +-
 .../apex/malhar/lib/util/ReusableStringReader.java |   2 +-
 .../apex/malhar/lib/util/ReversibleComparator.java |   2 +-
 .../malhar/lib/util/StorageAgentKeyValueStore.java |   4 +-
 .../org/apache/apex/malhar/lib/util/TableInfo.java |   2 +-
 .../apache/apex/malhar/lib/util/TimeBucketKey.java |   2 +-
 .../org/apache/apex/malhar/lib/util/TopNSort.java  |   2 +-
 .../apex/malhar/lib/util/TopNUniqueSort.java       |   2 +-
 .../lib/util/UnifierArrayHashMapFrequent.java      |   2 +-
 .../apex/malhar/lib/util/UnifierArrayList.java     |   2 +-
 .../apex/malhar/lib/util/UnifierBooleanAnd.java    |   2 +-
 .../apex/malhar/lib/util/UnifierBooleanOr.java     |   2 +-
 .../apex/malhar/lib/util/UnifierCountOccurKey.java |   2 +-
 .../apex/malhar/lib/util/UnifierHashMap.java       |   2 +-
 .../malhar/lib/util/UnifierHashMapFrequent.java    |   2 +-
 .../malhar/lib/util/UnifierHashMapInteger.java     |   2 +-
 .../apex/malhar/lib/util/UnifierHashMapRange.java  |   2 +-
 .../malhar/lib/util/UnifierHashMapSumKeys.java     |   2 +-
 .../apex/malhar/lib/util/UnifierKeyValRange.java   |   2 +-
 .../apache/apex/malhar/lib/util/UnifierMap.java    |   2 +-
 .../apache/apex/malhar/lib/util/UnifierRange.java  |   2 +-
 .../apex/malhar/lib/util/UnifierSumNumber.java     |   2 +-
 .../lib/util/comparator/SliceComparator.java       |   2 +-
 .../apache/apex/malhar/lib/util/package-info.java  |   2 +-
 .../apex/malhar/lib/util/time/WindowUtils.java     |   2 +-
 .../apex/malhar/lib/wal/FSWindowDataManager.java   |   4 +-
 .../apex/malhar/lib/wal/WindowDataManager.java     |   3 +-
 .../lib/window/accumulation/AbstractPojoJoin.java  |   5 +-
 .../lib/window/accumulation/PojoFullOuterJoin.java |   5 +-
 .../lib/window/accumulation/PojoInnerJoin.java     |   6 +-
 .../lib/window/accumulation/PojoLeftOuterJoin.java |   5 +-
 .../window/accumulation/PojoRightOuterJoin.java    |   5 +-
 .../malhar/lib/window/accumulation/TopNByKey.java  |   2 +-
 .../impl/KeyedWindowedMergeOperatorImpl.java       |   2 +-
 .../lib/window/impl/KeyedWindowedOperatorImpl.java |   2 +-
 .../window/impl/WindowedMergeOperatorFeatures.java |   3 +-
 .../malhar/lib/xml/AbstractXmlDOMOperator.java     |   2 +-
 .../org/apache/hadoop/io/file/tfile/DTFile.java    |   7 +-
 .../apex/malhar/lib/algo/BottomNMapTest.java       |   6 +-
 .../apex/malhar/lib/algo/BottomNUnifierTest.java   |   4 +-
 .../apex/malhar/lib/algo/BottomNUniqueMapTest.java |   6 +-
 .../malhar/lib/algo/CompareExceptCountMapTest.java |   6 +-
 .../apache/apex/malhar/lib/algo/DistinctTest.java  |   8 +-
 .../apex/malhar/lib/algo/FilterValuesTest.java     |   6 +-
 .../apex/malhar/lib/algo/InsertSortTest.java       |   6 +-
 .../apex/malhar/lib/algo/LeastFrequentKeyTest.java |   6 +-
 .../apex/malhar/lib/algo/MatchAllMapTest.java      |   6 +-
 .../apex/malhar/lib/algo/MatchAnyMapTest.java      |   6 +-
 .../apex/malhar/lib/algo/MergeSortNumberTest.java  |   6 +-
 .../apex/malhar/lib/algo/MostFrequentKeyTest.java  |   6 +-
 .../org/apache/apex/malhar/lib/algo/TopNTest.java  |   6 +-
 .../apex/malhar/lib/algo/TopNUniqueTest.java       |   6 +-
 .../apex/malhar/lib/algo/UniqueCounterTest.java    |   6 +-
 .../apex/malhar/lib/algo/UniqueValueCountTest.java |  12 +-
 .../apex/malhar/lib/appdata/ThreadUtils.java       |   2 +-
 .../datastructs/CacheLRUSynchronousFlushTest.java  |   6 +-
 .../appdata/datastructs/DimensionalTableTest.java  |   6 +-
 .../dimensions/CustomTimeBucketRegistryTest.java   |   7 +-
 .../appdata/dimensions/DimensionsEventTest.java    |   9 +-
 .../malhar/lib/appdata/gpo/GPOMutableTest.java     |   8 +-
 .../apex/malhar/lib/appdata/gpo/GPOUtilsTest.java  |   9 +-
 .../lib/appdata/gpo/SerdeFieldsDescriptorTest.java |   7 +-
 .../lib/appdata/gpo/SerdeListGPOMutableTest.java   |   7 +-
 .../lib/appdata/gpo/SerdeListPrimitiveTest.java    |   2 +-
 .../lib/appdata/gpo/SerdeListStringTest.java       |   2 +-
 .../apex/malhar/lib/appdata/query/MockQuery.java   |   4 +-
 .../apex/malhar/lib/appdata/query/MockResult.java  |  10 +-
 .../lib/appdata/query/MockResultSerializer.java    |  10 +-
 .../query/QueryManagerAsynchronousTest.java        |  10 +-
 .../appdata/query/QueryManagerSynchronousTest.java |   8 +-
 .../malhar/lib/appdata/query/QueueUtilsTest.java   |   6 +-
 .../query/SimpleDoneQueryQueueManagerTest.java     |   7 +-
 .../appdata/query/SimpleQueryQueueManagerTest.java |   4 +-
 .../lib/appdata/query/WEQueryQueueManagerTest.java |   5 +-
 .../appdata/query/WindowBoundedServiceTest.java    |   4 +-
 .../serde/MessageDeserializerFactoryTest.java      |   6 +-
 .../query/serde/SimpleDataValidatorTest.java       |   4 +-
 .../lib/appdata/schemas/CustomTimeBucketTest.java  |   2 +-
 .../schemas/DataQuerySnapshotDeserializerTest.java |   8 +-
 .../lib/appdata/schemas/DataQuerySnapshotTest.java |   2 +-
 .../schemas/DataResultSnapshotSerializerTest.java  |   8 +-
 .../DimensionalConfigurationSchemaTest.java        |   9 +-
 .../lib/appdata/schemas/DimensionalSchemaTest.java |   5 +-
 .../lib/appdata/schemas/FieldsDescriptorTest.java  |   2 +-
 .../lib/appdata/schemas/ResultFormatterTest.java   |   4 +-
 .../lib/appdata/schemas/SchemaQueryTest.java       |   6 +-
 .../schemas/SchemaRegistryMultipleTest.java        |   6 +-
 .../appdata/schemas/SchemaRegistrySingleTest.java  |   4 +-
 .../lib/appdata/schemas/SchemaTestUtils.java       |   2 +-
 .../lib/appdata/schemas/SchemaUtilsTest.java       |   2 +-
 .../lib/appdata/schemas/SnapshotSchemaTest.java    |   2 +-
 .../malhar/lib/appdata/schemas/TimeBucketTest.java |   4 +-
 .../snapshot/AppDataSnapshotServerMapTest.java     |   8 +-
 .../snapshot/AppDataSnapshotServerPojoTest.java    |  14 +-
 .../AppDataSnapshotServerTagsSupportTest.java      |  12 +-
 .../malhar/lib/bandwidth/BandwidthManagerTest.java |   5 +-
 .../lib/bandwidth/BandwidthPartitionerTest.java    |   2 +-
 .../codec/JavaSerializationStreamCodecTest.java    |   4 +-
 .../apex/malhar/lib/codec/KryoStreamCodecTest.java |   2 +-
 .../converter/ByteArrayToStringConverterTest.java  |   6 +-
 .../converter/MapToKeyValuePairConverterTest.java  |   8 +-
 .../StringValueToNumberConverterForMapTest.java    |   6 +-
 .../malhar/lib/counters/BasicCountersTest.java     |   2 +-
 .../malhar/lib/db/KeyValueStoreOperatorTest.java   |   4 +-
 .../TransactionableKeyValueStoreOperatorTest.java  |   4 +-
 .../apex/malhar/lib/db/cache/CacheManagerTest.java |   2 +-
 .../apex/malhar/lib/db/cache/CacheStoreTest.java   |   3 +-
 .../db/jdbc/JDBCLookupCacheBackedOperatorTest.java |   7 +-
 .../apache/apex/malhar/lib/db/jdbc/JdbcIOApp.java  |   6 +-
 .../apex/malhar/lib/db/jdbc/JdbcIOAppTest.java     |   2 +-
 .../db/jdbc/JdbcInputOperatorApplicationTest.java  |   4 +-
 ...dbcNonTransactionalBatchOutputOperatorTest.java |  17 +-
 .../JdbcNonTransactionalOutputOperatorTest.java    |   4 +-
 .../lib/db/jdbc/JdbcNonTransactionalStoreTest.java |   2 +-
 .../apex/malhar/lib/db/jdbc/JdbcOperatorTest.java  |   2 +-
 .../db/jdbc/JdbcPojoOperatorApplicationTest.java   |   2 +-
 .../malhar/lib/db/jdbc/JdbcPojoOperatorTest.java   |  16 +-
 .../lib/db/jdbc/JdbcPojoPollableOpeartorTest.java  |  14 +-
 .../apex/malhar/lib/db/jdbc/JdbcStoreTest.java     |   2 +-
 .../RedshiftJdbcTransactionalOperatorTest.java     |   4 +-
 .../lib/dedup/DeduperBoundedPOJOImplTest.java      |   8 +-
 .../apex/malhar/lib/dedup/DeduperOrderingTest.java |   4 +-
 .../malhar/lib/dedup/DeduperPartitioningTest.java  |   2 +-
 .../lib/dedup/DeduperTimeBasedPOJOImplTest.java    |  12 +-
 .../dimensions/CustomTimeBucketRegistryTest.java   |   4 +-
 .../lib/dimensions/DimensionsDescriptorTest.java   |  10 +-
 .../apex/malhar/lib/filter/FilterAppTest.java      |   5 +-
 .../apache/apex/malhar/lib/filter/FilterTest.java  |   5 +-
 .../malhar/lib/formatter/JsonFormatterTest.java    |  13 +-
 .../malhar/lib/formatter/XmlFormatterTest.java     |   8 +-
 .../lib/fs/FSRecordCompactionOperatorTest.java     |   4 +-
 .../lib/fs/GenericFileOutputOperatorTest.java      |   2 +-
 .../lib/fs/s3/S3DelimitedRecordReaderTest.java     |   9 +-
 .../lib/fs/s3/S3FixedWidthRecordReaderTest.java    |   9 +-
 .../fs/s3/S3InitiateFileUploadOperatorTest.java    |   8 +-
 .../malhar/lib/fs/s3/S3OutputModuleMockTest.java   |   4 +-
 .../apex/malhar/lib/fs/s3/S3ReconcilerTest.java    |   4 +-
 .../malhar/lib/fs/s3/S3RecordReaderMockTest.java   |   6 +-
 .../lib/helper/OperatorContextTestHelper.java      |   2 +-
 .../lib/helper/SamplePubSubWebSocketServlet.java   |   2 +-
 .../apex/malhar/lib/helper/TestPortContext.java    |   2 +-
 .../apex/malhar/lib/io/ApacheRandomLogsTest.java   |   4 +-
 .../malhar/lib/io/FTPStringInputOperatorTest.java  |  10 +-
 .../apex/malhar/lib/io/HttpGetMapOperatorTest.java |   6 +-
 .../lib/io/HttpJsonChunksInputOperatorTest.java    |   5 +-
 .../malhar/lib/io/HttpLinesInputOperatorTest.java  |   7 +-
 .../lib/io/HttpMultiValuedMapGetOperatorTest.java  |   8 +-
 .../malhar/lib/io/HttpPostOutputOperatorTest.java  |   4 +-
 .../lib/io/IdempotentStorageManagerTest.java       |   4 +-
 .../lib/io/PubSubWebSocketAppDataOperatorTest.java |   2 +-
 .../lib/io/PubSubWebSocketAppDataQueryTest.java    |   4 +-
 .../lib/io/PubSubWebSocketAppDataResultTest.java   |   2 +-
 .../malhar/lib/io/PubSubWebSocketOperatorTest.java |   9 +-
 .../apex/malhar/lib/io/SmtpOutputOperatorTest.java |   2 +-
 .../malhar/lib/io/SocketInputOperatorTest.java     |   6 +-
 .../lib/io/WebSocketServerInputOperatorTest.java   |   2 +-
 .../lib/io/block/AbstractBlockReaderTest.java      |   6 +-
 .../apex/malhar/lib/io/block/FSLineReaderTest.java |   7 +-
 .../malhar/lib/io/block/FSSliceReaderTest.java     |   6 +-
 .../malhar/lib/io/block/PartFileWriterTest.java    |   5 +-
 .../lib/io/block/ReadAheadLineReaderTest.java      |   2 +-
 ...stractFileInputOperatorFailureHandlingTest.java |  10 +-
 .../lib/io/fs/AbstractFileInputOperatorTest.java   |  15 +-
 .../lib/io/fs/AbstractFileOutputOperatorTest.java  |  10 +-
 .../malhar/lib/io/fs/AbstractReconcilerTest.java   |   6 +-
 .../fs/AbstractSingleFileOutputOperatorTest.java   |   8 +-
 .../AbstractThroughputFileInputOperatorTest.java   |   8 +-
 .../fs/AbstractWindowFileOutputOperatorTest.java   |   9 +-
 .../apex/malhar/lib/io/fs/BlockWriterTest.java     |   8 +-
 .../malhar/lib/io/fs/FSInputModuleAppTest.java     |  11 +-
 .../apex/malhar/lib/io/fs/FSInputModuleTest.java   |   6 +-
 .../lib/io/fs/FastMergerDecisionMakerTest.java     |   9 +-
 .../apex/malhar/lib/io/fs/FileMergerTest.java      |  18 +-
 .../malhar/lib/io/fs/FileSplitterBaseTest.java     |  10 +-
 .../malhar/lib/io/fs/FileSplitterInputTest.java    |  12 +-
 .../apex/malhar/lib/io/fs/FileSplitterTest.java    |  10 +-
 .../apex/malhar/lib/io/fs/FileStitcherTest.java    |  18 +-
 .../malhar/lib/io/fs/S3InputModuleAppTest.java     |   8 +-
 .../apex/malhar/lib/io/fs/SynchronizerTest.java    |  12 +-
 .../malhar/lib/io/fs/TailFsInputOperatorTest.java  |   6 +-
 .../io/jms/FSPsuedoTransactionableStoreTest.java   |   2 +-
 .../lib/io/jms/JMSMultiPortOutputOperatorTest.java |   9 +-
 .../lib/io/jms/JMSObjectInputOperatorTest.java     |   7 +-
 .../malhar/lib/io/jms/JMSOutputOperatorTest.java   |   6 +-
 .../lib/io/jms/JMSStringInputOperatorTest.java     |   6 +-
 .../apache/apex/malhar/lib/io/jms/JMSTestBase.java |   2 +-
 .../lib/io/jms/JMSTransactionableStoreTest.java    |   2 +-
 .../io/jms/JMSTransactionableStoreTestBase.java    |  11 +-
 .../lib/io/jms/SQSStringInputOperatorTest.java     |  10 +-
 .../apache/apex/malhar/lib/io/jms/SQSTestBase.java |   2 +-
 .../apex/malhar/lib/join/AntiJoinOperatorTest.java |  12 +-
 .../malhar/lib/join/MapTimeBasedJoinOperator.java  |   9 +-
 .../malhar/lib/join/POJOInnerJoinOperatorTest.java |   8 +-
 .../lib/join/POJOPartitionJoinOperatorTest.java    |   2 +-
 .../lib/join/POJOTimeBasedJoinOperatorTest.java    |   8 +-
 .../apex/malhar/lib/join/SemiJoinOperatorTest.java |  12 +-
 .../logs/ApacheLogParseMapOutputOperatorTest.java  |   6 +-
 .../lib/logs/ApacheLogParseOperatorTest.java       |   6 +-
 .../lib/logs/FilteredLineToTokenArrayListTest.java |   6 +-
 .../lib/logs/FilteredLineToTokenHashMapTest.java   |   6 +-
 .../lib/logs/FilteredLineTokenizerKeyValTest.java  |   6 +-
 .../malhar/lib/logs/LineToTokenArrayListTest.java  |   6 +-
 .../malhar/lib/logs/LineToTokenHashMapTest.java    |   6 +-
 .../malhar/lib/logs/LineTokenizerKeyValTest.java   |   6 +-
 .../apex/malhar/lib/logs/LineTokenizerTest.java    |   6 +-
 .../logs/MultiWindowDimensionAggregationTest.java  |   7 +-
 .../malhar/lib/logs/RegexMatchMapOperatorTest.java |   6 +-
 .../malhar/lib/logs/TopNUniqueSiteStatsTest.java   |   7 +-
 .../apex/malhar/lib/math/AverageKeyValTest.java    |   8 +-
 .../apache/apex/malhar/lib/math/AverageTest.java   |   8 +-
 .../apex/malhar/lib/math/CountOccuranceTest.java   |   6 +-
 .../apache/apex/malhar/lib/math/DivisionTest.java  |   6 +-
 .../apex/malhar/lib/math/LogicalCompareTest.java   |   7 +-
 .../lib/math/LogicalCompareToConstantTest.java     |   6 +-
 .../apex/malhar/lib/math/MarginKeyValTest.java     |   8 +-
 .../apache/apex/malhar/lib/math/MarginMapTest.java |   6 +-
 .../apache/apex/malhar/lib/math/MarginTest.java    |   6 +-
 .../apache/apex/malhar/lib/math/MaxKeyValTest.java |   9 +-
 .../org/apache/apex/malhar/lib/math/MaxTest.java   |   6 +-
 .../apache/apex/malhar/lib/math/MinKeyValTest.java |   9 +-
 .../org/apache/apex/malhar/lib/math/MinTest.java   |   6 +-
 .../malhar/lib/math/MultiplyByConstantTest.java    |   6 +-
 .../apex/malhar/lib/math/RangeKeyValTest.java      |  10 +-
 .../org/apache/apex/malhar/lib/math/RangeTest.java |   8 +-
 .../apex/malhar/lib/math/RunningAverageTest.java   |   4 +-
 .../org/apache/apex/malhar/lib/math/SigmaTest.java |   6 +-
 .../apex/malhar/lib/math/SquareCalculusTest.java   |   6 +-
 .../apache/apex/malhar/lib/math/SumKeyValTest.java |   8 +-
 .../org/apache/apex/malhar/lib/math/SumTest.java   |   6 +-
 .../XmlKeyValueStringCartesianProductTest.java     |   2 +-
 .../multiwindow/MultiWindowRangeKeyValTest.java    |   8 +-
 .../lib/multiwindow/MultiWindowSumKeyValTest.java  |   8 +-
 .../lib/multiwindow/SimpleMovingAverageTest.java   |   8 +-
 .../malhar/lib/multiwindow/SlidingWindowTest.java  |   8 +-
 .../lib/multiwindow/SortedMovingWindowTest.java    |   5 +-
 .../lib/parser/XmlParserApplicationTest.java       |  12 +-
 .../apex/malhar/lib/parser/XmlParserTest.java      |   8 +-
 .../lib/partitioner/StatelessPartitionerTest.java  |   5 +-
 .../StatelessThroughputBasedPartitionerTest.java   |   5 +-
 .../apex/malhar/lib/projection/ActivateTest.java   |   2 +-
 .../apex/malhar/lib/projection/ProjectionTest.java |   4 +-
 .../malhar/lib/script/JavaScriptOperatorTest.java  |   8 +-
 .../lib/state/managed/BucketsFileSystemTest.java   |   5 +-
 .../lib/state/managed/DefaultBucketTest.java       |   6 +-
 .../managed/IncrementalCheckpointManagerTest.java  |   6 +-
 .../lib/state/managed/ManagedStateImplTest.java    |   7 +-
 .../lib/state/managed/ManagedStateTestUtils.java   |   6 +-
 .../state/managed/ManagedTimeStateImplTest.java    |   8 +-
 .../managed/ManagedTimeUnifiedStateImplTest.java   |   4 +-
 .../lib/state/managed/MockManagedStateContext.java |   7 +-
 .../MovingBoundaryTimeBucketAssignerTest.java      |   3 +-
 .../malhar/lib/state/managed/StateTrackerTest.java |   5 +-
 ...SequentialSpillableIdentifierGeneratorTest.java |   2 +-
 .../spillable/SpillableArrayListImplTest.java      |   7 +-
 .../SpillableArrayListMultimapImplTest.java        |   4 +-
 .../lib/state/spillable/SpillableMapImplTest.java  |   4 +-
 .../spillable/SpillableSetMultimapImplTest.java    |   7 +-
 .../lib/state/spillable/SpillableTestUtils.java    |   4 +-
 .../inmem/InMemorySpillableStateStoreTest.java     |   2 +-
 .../lib/statistics/MeridianOperatorTest.java       |   8 +-
 .../malhar/lib/statistics/ModeOperatorTest.java    |   8 +-
 .../statistics/StandardDeviationOperatorTest.java  |   6 +-
 .../lib/statistics/WeightedMeanOperatorTest.java   |   8 +-
 .../malhar/lib/stream/ArrayListAggregatorTest.java |   6 +-
 .../malhar/lib/stream/ArrayListToItemTest.java     |   6 +-
 .../malhar/lib/stream/ConsolidatorKeyValTest.java  |   8 +-
 .../apache/apex/malhar/lib/stream/CounterTest.java |   6 +-
 .../apex/malhar/lib/stream/DevNullCounterTest.java |   6 +-
 .../apache/apex/malhar/lib/stream/DevNullTest.java |   6 +-
 .../malhar/lib/stream/HashMapToKeyValPairTest.java |   6 +-
 .../lib/stream/JsonByteArrayOperatorTest.java      |   4 +-
 .../malhar/lib/stream/KeyPairToHashMapTest.java    |   8 +-
 .../malhar/lib/stream/RoundRobinHashMapTest.java   |   6 +-
 .../malhar/lib/stream/StreamDuplicaterTest.java    |   6 +-
 .../apex/malhar/lib/stream/StreamMergerTest.java   |   6 +-
 .../lib/testbench/ActiveMQMessageGenerator.java    |   2 +-
 .../malhar/lib/testbench/EventClassifierTest.java  |   4 +-
 .../malhar/lib/testbench/EventGeneratorTest.java   |   4 +-
 .../malhar/lib/testbench/EventIncrementerTest.java |   4 +-
 .../lib/testbench/FilteredEventClassifierTest.java |   4 +-
 .../lib/testbench/RandomEventGeneratorTest.java    |   4 +-
 .../lib/testbench/RandomWordGeneratorTest.java     |   2 +-
 .../lib/testbench/SeedEventClassifierTest.java     |   4 +-
 .../lib/testbench/SeedEventGeneratorTest.java      |   4 +-
 .../lib/testbench/ThroughputCounterTest.java       |   4 +-
 .../lib/transform/TransformOperatorAppTest.java    |   5 +-
 .../lib/transform/TransformOperatorTest.java       |   7 +-
 .../util/DimensionTimeBucketSumOperatorTest.java   |   5 +-
 .../lib/util/JavaScriptFilterOperatorTest.java     |   6 +-
 .../apex/malhar/lib/util/KryoCloneUtilsTest.java   |   2 +-
 .../apache/apex/malhar/lib/util/PojoUtilsTest.java |  78 +++---
 .../apex/malhar/lib/util/TestObjAllTypes.java      |   2 +-
 .../org/apache/apex/malhar/lib/util/TestUtils.java |   2 +-
 .../apex/malhar/lib/util/time/WindowUtilsTest.java |   4 +-
 .../malhar/lib/wal/FSWindowDataManagerTest.java    |   5 +-
 .../apex/malhar/lib/wal/FileSystemWALTest.java     |   2 +-
 .../lib/window/SpillableWindowedStorageTest.java   |   2 +-
 .../malhar/lib/window/WindowedOperatorTest.java    |   4 +-
 .../lib/window/accumulation/PojoInnerJoinTest.java |   4 +-
 .../lib/window/accumulation/PojoOuterJoinTest.java |   4 +-
 .../lib/window/accumulation/TopNByKeyTest.java     |   2 +-
 .../KeyedWindowedMergeOperatorTestApplication.java |   4 +-
 .../lib/window/impl/WindowedMergeOperatorTest.java |   5 +-
 .../impl/WindowedMergeOperatorTestApplication.java |   2 +-
 .../malhar/lib/window/sample/pi/Application.java   |   2 +-
 .../lib/window/sample/wordcount/Application.java   |   4 +-
 pom.xml                                            |   4 +-
 samples/pom.xml                                    |  55 -----
 .../samples/lib/algo/AllAfterMatchMapSample.java   |  65 -----
 .../samples/lib/algo/BottomNMapSample.java         |  65 -----
 .../datatorrent/samples/lib/algo/package-info.java |  25 --
 .../samples/lib/io/HttpInputOperatorSample.java    |  64 -----
 .../datatorrent/samples/lib/io/package-info.java   |  24 --
 .../samples/lib/math/AverageKeyValSample.java      |  64 -----
 .../samples/lib/math/AverageSample.java            |  70 ------
 .../datatorrent/samples/lib/math/ChangeSample.java |  70 ------
 .../samples/lib/math/CompreMapSample.java          |  61 -----
 .../samples/lib/math/CountKeyValSample.java        |  62 -----
 .../samples/lib/math/CountKeyValues.java           |  88 -------
 .../samples/lib/math/LogicalCompareSample.java     |  62 -----
 .../samples/lib/math/PartitionMathSumSample.java   |  71 ------
 .../lib/math/RandomEvenetGeneratorSample.java      |  57 -----
 .../samples/lib/math/RandomKeyValMap.java          |  84 -------
 .../samples/lib/math/RandomKeyValues.java          |  84 -------
 .../samples/lib/math/RandomPairIntegers.java       |  84 -------
 .../datatorrent/samples/lib/math/ScriptSample.java |  61 -----
 .../samples/lib/math/SingleKeyValMap.java          |  82 -------
 .../datatorrent/samples/lib/math/package-info.java |  25 --
 .../java/com/datatorrent/samples/package-info.java |  25 --
 .../lib/math/PartitionMathSumSampleTest.java       |  44 ----
 samples/src/test/resources/log4j.properties        |  43 ----
 sql/pom.xml                                        |   2 +-
 .../malhar/sql/codegen/ExpressionCompiler.java     |   2 +-
 .../sql/operators/FilterTransformOperator.java     |   6 +-
 .../apex/malhar/sql/operators/LineReader.java      |   2 +-
 .../apex/malhar/sql/planner/ApexRelNode.java       |   2 +-
 .../apex/malhar/sql/table/CSVMessageFormat.java    |   6 +-
 .../apache/apex/malhar/sql/KafkaEndpointTest.java  |   4 +-
 .../java/org/apache/apex/malhar/sql/SerDeTest.java |   4 +-
 .../apache/apex/malhar/sql/StreamEndpointTest.java |   4 +-
 stream/pom.xml                                     |   2 +-
 .../apex/malhar/stream/api/WindowedStream.java     |   3 +-
 .../malhar/stream/api/impl/ApexStreamImpl.java     |   2 +-
 .../stream/api/impl/ApexWindowedStreamImpl.java    |   3 +-
 .../apex/malhar/stream/api/impl/StreamFactory.java |   2 +-
 .../stream/sample/ApplicationWithStreamAPI.java    |   2 +-
 .../sample/LocalTestWithoutStreamApplication.java  |   3 +-
 .../apex/malhar/stream/sample/MyStreamTest.java    |   3 +-
 .../stream/sample/WordCountWithStreamAPI.java      |   2 +-
 1576 files changed, 6492 insertions(+), 7802 deletions(-)

diff --git a/apps/filecopy/pom.xml b/apps/filecopy/pom.xml
index d6a0ccd..69b2e95 100644
--- a/apps/filecopy/pom.xml
+++ b/apps/filecopy/pom.xml
@@ -24,7 +24,7 @@
   <parent>
     <artifactId>malhar-apps</artifactId>
     <groupId>org.apache.apex</groupId>
-    <version>3.9.0-SNAPSHOT</version>
+    <version>4.0.0-SNAPSHOT</version>
   </parent>
 
   <artifactId>filecopy</artifactId>
@@ -79,7 +79,7 @@
     <dependency>
       <groupId>org.apache.apex</groupId>
       <artifactId>malhar-library</artifactId>
-      <version>3.9.0-SNAPSHOT</version>
+      <version>4.0.0-SNAPSHOT</version>
       <exclusions>
         <exclusion>
           <groupId>*</groupId>
diff --git a/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java b/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java
index 2688f1c..bb1e1d9 100644
--- a/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java
+++ b/apps/filecopy/src/main/java/com/datatorrent/apps/copy/HDFSFileCopyApp.java
@@ -19,14 +19,14 @@
 
 package com.datatorrent.apps.copy;
 
+import org.apache.apex.malhar.lib.io.fs.FSInputModule;
+import org.apache.apex.malhar.lib.io.fs.HDFSFileCopyModule;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.io.fs.FSInputModule;
-import com.datatorrent.lib.io.fs.HDFSFileCopyModule;
 
 /**
  * Application for HDFS to HDFS file copy
diff --git a/apps/logstream/pom.xml b/apps/logstream/pom.xml
index 07a5eee..13cc865 100644
--- a/apps/logstream/pom.xml
+++ b/apps/logstream/pom.xml
@@ -24,7 +24,7 @@
   <parent>
     <artifactId>malhar-apps</artifactId>
     <groupId>org.apache.apex</groupId>
-    <version>3.9.0-SNAPSHOT</version>
+    <version>4.0.0-SNAPSHOT</version>
   </parent>
 
   <groupId>org.apache.apex</groupId>
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java
index 495987e..74997f4 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/AggregationsToRedisOperator.java
@@ -22,8 +22,8 @@ import java.util.ArrayList;
 import java.util.HashMap;
 import java.util.Map;
 
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.util.KeyValPair;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.util.KeyValPair;
 
 import com.datatorrent.common.util.BaseOperator;
 import com.datatorrent.api.DefaultInputPort;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java
index 82c9214..e3858b4 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application.java
@@ -28,25 +28,26 @@ import java.util.Set;
 import org.apache.apex.malhar.contrib.misc.streamquery.SelectOperator;
 import org.apache.apex.malhar.contrib.misc.streamquery.condition.EqualValueCondition;
 import org.apache.apex.malhar.lib.utils.PubSubHelper;
+import org.apache.apex.malhar.contrib.redis.RedisKeyValPairOutputOperator;
+import org.apache.apex.malhar.contrib.redis.RedisMapOutputOperator;
+import org.apache.apex.malhar.contrib.redis.RedisNumberSummationMapOutputOperator;
+import org.apache.apex.malhar.lib.algo.TopN;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.io.PubSubWebSocketOutputOperator;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.logs.MultiWindowDimensionAggregation;
+import org.apache.apex.malhar.lib.logs.MultiWindowDimensionAggregation.AggregateOperation;
+import org.apache.apex.malhar.lib.stream.Counter;
+import org.apache.apex.malhar.lib.stream.JsonByteArrayOperator;
+import org.apache.apex.malhar.lib.streamquery.index.ColumnIndex;
+import org.apache.apex.malhar.lib.util.AbstractDimensionTimeBucketOperator;
+import org.apache.apex.malhar.lib.util.DimensionTimeBucketSumOperator;
+import org.apache.commons.lang.StringUtils;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.Operator.InputPort;
 import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.contrib.redis.RedisKeyValPairOutputOperator;
-import com.datatorrent.contrib.redis.RedisMapOutputOperator;
-import com.datatorrent.contrib.redis.RedisNumberSummationMapOutputOperator;
-import com.datatorrent.lib.algo.TopN;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.io.PubSubWebSocketOutputOperator;
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.logs.MultiWindowDimensionAggregation;
-import com.datatorrent.lib.logs.MultiWindowDimensionAggregation.AggregateOperation;
-import com.datatorrent.lib.stream.Counter;
-import com.datatorrent.lib.stream.JsonByteArrayOperator;
-import com.datatorrent.lib.streamquery.index.ColumnIndex;
-import com.datatorrent.lib.util.AbstractDimensionTimeBucketOperator;
-import com.datatorrent.lib.util.DimensionTimeBucketSumOperator;
 
 /**
  * Log stream processing application based on Apex platform.<br>
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java
index 333b877..a458b3c 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/Application1.java
@@ -18,11 +18,10 @@
  */
 package com.datatorrent.apps.logstream;
 
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.stream.JsonByteArrayOperator;
 import org.apache.hadoop.conf.Configuration;
 
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.stream.JsonByteArrayOperator;
-
 import com.datatorrent.api.Context.PortContext;
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.StreamingApplication;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java
index 2650795..de96781 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperator.java
@@ -28,10 +28,11 @@ import com.google.common.collect.Sets;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.codec.KryoSerializableStreamCodec;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
 import org.apache.commons.lang.mutable.MutableDouble;
 
-import com.datatorrent.lib.codec.KryoSerializableStreamCodec;
-import com.datatorrent.lib.logs.DimensionObject;
 import com.datatorrent.api.*;
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.Partitioner.PartitioningContext;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java
index e495ee6..606edb7 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/DimensionOperatorUnifier.java
@@ -26,10 +26,9 @@ import java.util.Map.Entry;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.logs.DimensionObject;
 import org.apache.commons.lang.mutable.MutableDouble;
 
-import com.datatorrent.lib.logs.DimensionObject;
-
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.Operator.Unifier;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java
index 57c558b..74dc3a6 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogScoreOperator.java
@@ -20,10 +20,12 @@ package com.datatorrent.apps.logstream;
 
 import com.datatorrent.common.util.BaseOperator;
 import com.datatorrent.api.DefaultInputPort;
-import com.datatorrent.lib.logs.DimensionObject;
+
 import java.util.HashMap;
 import java.util.Map;
 
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+
 /**
  * Log Score
  *
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java
index a09e64f..8df62fa 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamTopN.java
@@ -28,9 +28,9 @@ import com.google.common.collect.Sets;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.datatorrent.lib.algo.TopN;
-import com.datatorrent.lib.codec.KryoSerializableStreamCodec;
-import com.datatorrent.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.algo.TopN;
+import org.apache.apex.malhar.lib.codec.KryoSerializableStreamCodec;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
 
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultPartition;
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java
index 29c92e6..715cb6b 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/LogstreamWidgetOutputOperator.java
@@ -26,11 +26,10 @@ import java.util.Map.Entry;
 
 import javax.validation.constraints.NotNull;
 
+import org.apache.apex.malhar.lib.io.WidgetOutputOperator;
+import org.apache.apex.malhar.lib.logs.DimensionObject;
 import org.apache.commons.lang3.tuple.MutablePair;
 
-import com.datatorrent.lib.io.WidgetOutputOperator;
-import com.datatorrent.lib.logs.DimensionObject;
-
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultInputPort;
 
diff --git a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java
index a72dd05..a4e7fd2 100644
--- a/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java
+++ b/apps/logstream/src/main/java/com/datatorrent/apps/logstream/RabbitMQLogsInputOperator.java
@@ -24,13 +24,14 @@ import org.codehaus.jettison.json.JSONObject;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.contrib.rabbitmq.AbstractSinglePortRabbitMQInputOperator;
+
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.Partitioner.PartitioningContext;
 import com.datatorrent.api.DefaultPartition;
 import com.datatorrent.api.Partitioner;
 import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.contrib.rabbitmq.AbstractSinglePortRabbitMQInputOperator;
 
 /**
  *
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java
index 1ecd094..6eadc6e 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorTest.java
@@ -26,11 +26,10 @@ import java.util.Map.Entry;
 import org.junit.Assert;
 import org.junit.Test;
 
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
 import org.apache.commons.lang.mutable.MutableDouble;
 
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.testbench.CollectorTestSink;
-
 import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
 import com.datatorrent.netlet.util.DTThrowable;
 
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java
index a835dab..05929e2 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/DimensionOperatorUnifierTest.java
@@ -26,11 +26,10 @@ import org.junit.Assert;
 
 import org.junit.Test;
 
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
 import org.apache.commons.lang.mutable.MutableDouble;
 
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.testbench.CollectorTestSink;
-
 /**
  *
  * Tests logstream dimension operator unifier.
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java
index 2b7a5bd..a2bc229 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/FilterOperatorTest.java
@@ -23,7 +23,7 @@ import org.junit.Assert;
 
 import org.junit.Test;
 
-import com.datatorrent.lib.testbench.CollectorTestSink;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
 
 import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
 
diff --git a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java
index 69d0567..d83373c 100644
--- a/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java
+++ b/apps/logstream/src/test/java/com/datatorrent/apps/logstream/LogstreamTopNTest.java
@@ -26,11 +26,10 @@ import java.util.Map;
 import org.junit.Assert;
 import org.junit.Test;
 
+import org.apache.apex.malhar.lib.logs.DimensionObject;
+import org.apache.apex.malhar.lib.testbench.CollectorTestSink;
 import org.apache.commons.lang.mutable.MutableDouble;
 
-import com.datatorrent.lib.logs.DimensionObject;
-import com.datatorrent.lib.testbench.CollectorTestSink;
-
 import com.datatorrent.apps.logstream.PropertyRegistry.LogstreamPropertyRegistry;
 
 /**
diff --git a/apps/pom.xml b/apps/pom.xml
index 679d623..1f46ecd 100644
--- a/apps/pom.xml
+++ b/apps/pom.xml
@@ -25,7 +25,7 @@
   <parent>
     <groupId>org.apache.apex</groupId>
     <artifactId>malhar</artifactId>
-    <version>3.9.0-SNAPSHOT</version>
+    <version>4.0.0-SNAPSHOT</version>
   </parent>
 
   <artifactId>malhar-apps</artifactId>
diff --git a/benchmark/pom.xml b/benchmark/pom.xml
index 0822549..1b921d7 100644
--- a/benchmark/pom.xml
+++ b/benchmark/pom.xml
@@ -24,7 +24,7 @@
   <parent>
     <artifactId>malhar</artifactId>
     <groupId>org.apache.apex</groupId>
-    <version>3.9.0-SNAPSHOT</version>
+    <version>4.0.0-SNAPSHOT</version>
   </parent>
 
   <groupId>org.apache.apex</groupId>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java b/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java
index aa10eea..1491f47 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/ApplicationFixed.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import org.apache.hadoop.conf.Configuration;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java
index d8d51b8..6992bd8 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/Benchmark.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import org.apache.hadoop.conf.Configuration;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java
index bf5b876..6e29652 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppInput.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import org.apache.hadoop.conf.Configuration;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java
index 4f12791..95eaacc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseAppOutput.java
@@ -16,8 +16,9 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
@@ -25,7 +26,6 @@ import com.datatorrent.api.DAG.Locality;
 
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
 
 /**
  *
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java
index 8ae0a94..12e43e2 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseInputOperator.java
@@ -16,15 +16,15 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import java.util.ArrayList;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.datatorrent.contrib.couchbase.AbstractCouchBaseInputOperator;
-import com.datatorrent.contrib.couchbase.CouchBaseWindowStore;
+import org.apache.apex.malhar.contrib.couchbase.AbstractCouchBaseInputOperator;
+import org.apache.apex.malhar.contrib.couchbase.CouchBaseWindowStore;
 
 /**
  * <p>CouchBaseInputOperator class.</p>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java
index 8b6fc49..43a8e91 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/CouchBaseOutputOperator.java
@@ -16,9 +16,9 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
-import com.datatorrent.contrib.couchbase.AbstractCouchBaseSetOperator;
+import org.apache.apex.malhar.contrib.couchbase.AbstractCouchBaseSetOperator;
 
 /**
  * <p>CouchBaseOutputOperator class.</p>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java
index f2582bd..c3248bf 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/FixedTuplesInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import java.util.ArrayList;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java b/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java
index 3342771..d5af7a5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/RandomMapOutput.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import java.util.HashMap;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java b/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java
index 7d02de2..7224df3 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/RandomWordInputModule.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import javax.validation.constraints.Min;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java
index 8c55404..1138e84 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/WordCountOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 /*
  * To change this template, choose Tools | Templates and open the template in the editor.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java
index 0a880fd..3282e32 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputBenchmarkApplication.java
@@ -16,16 +16,16 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.aerospike;
+package org.apache.apex.benchmark.aerospike;
 
+import org.apache.apex.malhar.contrib.aerospike.AerospikeTransactionalStore;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.aerospike.AerospikeTransactionalStore;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
 
 /**
  * Application to benchmark the performance of aerospike output operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java
index f9ee689..bb3661b 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/aerospike/AerospikeOutputOperator.java
@@ -16,16 +16,16 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.aerospike;
+package org.apache.apex.benchmark.aerospike;
 
 import java.util.List;
 
+import org.apache.apex.malhar.contrib.aerospike.AbstractAerospikeTransactionalPutOperator;
+
 import com.aerospike.client.AerospikeException;
 import com.aerospike.client.Bin;
 import com.aerospike.client.Key;
 
-import com.datatorrent.contrib.aerospike.AbstractAerospikeTransactionalPutOperator;
-
 /**
  * <p>AerospikeOutputOperator class.</p>
  *
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java
index f74311e..8e1d8b8 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkApplication.java
@@ -16,8 +16,13 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.algo;
+package org.apache.apex.benchmark.algo;
 
+import org.apache.apex.malhar.lib.algo.UniqueCounter;
+import org.apache.apex.malhar.lib.converter.MapToKeyHashValuePairConverter;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.stream.Counter;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context;
@@ -27,13 +32,6 @@ import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
 import com.datatorrent.common.partitioner.StatelessPartitioner;
 
-import com.datatorrent.lib.algo.UniqueCounter;
-import com.datatorrent.lib.converter.MapToKeyHashValuePairConverter;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-
-import com.datatorrent.lib.stream.Counter;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
-
 /**
  * Application to demonstrate PartitionableUniqueCount operator. <br>
  * The input operator generate random keys, which is sent to
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java
index 46d503f..6f91cd1 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputBenchmarkApplication.java
@@ -16,8 +16,10 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.cassandra;
+package org.apache.apex.benchmark.cassandra;
 
+import org.apache.apex.malhar.contrib.cassandra.CassandraTransactionalStore;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
@@ -26,10 +28,6 @@ import com.datatorrent.api.StreamingApplication;
 
 import com.datatorrent.api.annotation.ApplicationAnnotation;
 
-import com.datatorrent.contrib.cassandra.CassandraTransactionalStore;
-
-import com.datatorrent.lib.testbench.RandomEventGenerator;
-
 /**
  * Application to benchmark the performance of cassandra output operator.
  * The operator was tested on following configuration:
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java
index 592d8a2..90bc44b 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/cassandra/CassandraOutputOperator.java
@@ -16,15 +16,15 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.cassandra;
+package org.apache.apex.benchmark.cassandra;
+
+import org.apache.apex.malhar.contrib.cassandra.AbstractCassandraTransactionableOutputOperator;
 
 import com.datastax.driver.core.BoundStatement;
 import com.datastax.driver.core.PreparedStatement;
 import com.datastax.driver.core.Statement;
 import com.datastax.driver.core.exceptions.DriverException;
 
-import com.datatorrent.contrib.cassandra.AbstractCassandraTransactionableOutputOperator;
-
 /**
  * <p>CassandraOutputOperator class.</p>
  *
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java
index ce0821c..56dd731 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSByteOutputOperator.java
@@ -16,13 +16,13 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.fs;
+package org.apache.apex.benchmark.fs;
 
 import java.util.Arrays;
 
 import javax.validation.constraints.Min;
 
-import com.datatorrent.lib.io.fs.AbstractFileOutputOperator;
+import org.apache.apex.malhar.lib.io.fs.AbstractFileOutputOperator;
 
 /**
  * This output operator receives
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java
index 7a63d18..baddbc1 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/fs/FSOutputOperatorBenchmark.java
@@ -16,8 +16,10 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.fs;
+package org.apache.apex.benchmark.fs;
 
+import org.apache.apex.malhar.lib.counters.BasicCounters;
+import org.apache.apex.malhar.lib.testbench.RandomWordGenerator;
 import org.apache.commons.lang.mutable.MutableLong;
 import org.apache.hadoop.conf.Configuration;
 
@@ -29,10 +31,6 @@ import com.datatorrent.api.StreamingApplication;
 
 import com.datatorrent.api.annotation.ApplicationAnnotation;
 
-import com.datatorrent.lib.counters.BasicCounters;
-
-import com.datatorrent.lib.testbench.RandomWordGenerator;
-
 /**
  * Application used to benchmark HDFS output operator
  * The DAG consists of random word generator operator that is
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java
index 95fa961..2ecb2eb 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkingApp.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
 
 import java.sql.SQLException;
 import java.sql.Statement;
@@ -28,6 +28,9 @@ import javax.validation.constraints.Min;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.hive.AbstractFSRollingOutputOperator;
+import org.apache.apex.malhar.hive.HiveOperator;
+import org.apache.apex.malhar.hive.HiveStore;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.OperatorContext;
@@ -39,9 +42,6 @@ import com.datatorrent.api.InputOperator;
 import com.datatorrent.api.StreamingApplication;
 
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.hive.AbstractFSRollingOutputOperator;
-import com.datatorrent.contrib.hive.HiveOperator;
-import com.datatorrent.contrib.hive.HiveStore;
 
 /**
  * Application used to benchmark HIVE Insert operator
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java
index 98d9ce3..800fa5a 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/hive/HiveMapInsertBenchmarkingApp.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
 
 import java.sql.SQLException;
 import java.sql.Statement;
@@ -27,18 +27,17 @@ import java.util.Map;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.benchmark.RandomMapOutput;
+import org.apache.apex.malhar.hive.AbstractFSRollingOutputOperator;
+import org.apache.apex.malhar.hive.HiveOperator;
+import org.apache.apex.malhar.hive.HiveStore;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.benchmark.RandomMapOutput;
-
-import com.datatorrent.contrib.hive.AbstractFSRollingOutputOperator;
-import com.datatorrent.contrib.hive.HiveOperator;
-import com.datatorrent.contrib.hive.HiveStore;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
 
 
 /**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java
index e147ad7..0bde5ed 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkKafkaInputOperator.java
@@ -16,10 +16,11 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
+
+import org.apache.apex.malhar.contrib.kafka.AbstractKafkaInputOperator;
 
 import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.contrib.kafka.AbstractKafkaInputOperator;
 
 import kafka.message.Message;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java
index 6353c37..37b4d0a 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/BenchmarkPartitionableKafkaOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
 
 import java.util.ArrayList;
 import java.util.Collection;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java
index ead6c66..e51da05 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmark.java
@@ -16,11 +16,13 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
 
 import java.util.Properties;
 
-
+import org.apache.apex.malhar.contrib.kafka.HighlevelKafkaConsumer;
+import org.apache.apex.malhar.contrib.kafka.KafkaConsumer;
+import org.apache.apex.malhar.contrib.kafka.SimpleKafkaConsumer;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.OperatorContext;
@@ -35,9 +37,6 @@ import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
 
 import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.contrib.kafka.HighlevelKafkaConsumer;
-import com.datatorrent.contrib.kafka.KafkaConsumer;
-import com.datatorrent.contrib.kafka.SimpleKafkaConsumer;
 
 /**
  * The stream app to test the benckmark of kafka
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java
index 0dd4352..e8578a6 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmark.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
 
 import org.apache.hadoop.conf.Configuration;
 
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java
index 65601d5..64ef2c7 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/kafka/KafkaTestPartitioner.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
 
 import kafka.producer.Partitioner;
 import kafka.utils.VerifiableProperties;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java b/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java
index 5ee1888..9f99b8f 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/package-info.java
@@ -19,4 +19,4 @@
 /**
  * Apex malhar performance demonstration application.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java b/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java
index b86cd01..065f546 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkApplication.java
@@ -16,8 +16,12 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.script;
+package org.apache.apex.benchmark.script;
 
+import org.apache.apex.benchmark.RandomMapOutput;
+import org.apache.apex.malhar.contrib.ruby.RubyOperator;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -27,12 +31,6 @@ import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
 
-import com.datatorrent.benchmark.RandomMapOutput;
-import com.datatorrent.contrib.ruby.RubyOperator;
-
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
-
 /**
  *
  * Application to benchmark the performance of ruby operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java
index 300ea64..4419a11 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkApp.java
@@ -16,8 +16,9 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
 
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
 import org.apache.apex.malhar.lib.state.spillable.managed.ManagedStateSpillableStateStore;
 import org.apache.hadoop.conf.Configuration;
 
@@ -26,7 +27,6 @@ import com.google.common.base.Preconditions;
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.fileaccess.TFileImpl;
 
 @ApplicationAnnotation(name = "SpillableBenchmarkApp")
 /**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java
index 9bee9a6..0285f04 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
 
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.InputOperator;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java
index 7c45106..c827acb 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/spillable/SpillableTestOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java
index 2dc6f0d..5ba8a33 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkApp.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.state;
+package org.apache.apex.benchmark.state;
 
 import java.io.IOException;
 import java.io.Serializable;
@@ -27,7 +27,9 @@ import org.joda.time.Duration;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
 import org.apache.apex.malhar.lib.state.managed.ManagedTimeUnifiedStateImpl;
+import org.apache.apex.malhar.lib.util.KeyValPair;
 import org.apache.hadoop.conf.Configuration;
 
 import com.google.common.collect.Lists;
@@ -43,8 +45,6 @@ import com.datatorrent.api.StatsListener;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
 import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.fileaccess.TFileImpl;
-import com.datatorrent.lib.util.KeyValPair;
 
 @ApplicationAnnotation(name = "ManagedStateBenchmark")
 /**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java
index 60a775c..eea4ac5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/state/StoreOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.state;
+package org.apache.apex.benchmark.state;
 
 import java.nio.ByteBuffer;
 import java.util.LinkedList;
@@ -28,6 +28,7 @@ import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
 import org.apache.apex.malhar.lib.state.managed.ManagedTimeUnifiedStateImpl;
+import org.apache.apex.malhar.lib.util.KeyValPair;
 
 import com.google.common.collect.Maps;
 
@@ -35,7 +36,6 @@ import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.api.Operator;
 import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.util.KeyValPair;
 import com.datatorrent.netlet.util.Slice;
 
 /**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java
index b0b7314..ef853e1 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/DevNullCounterBenchmark.java
@@ -16,8 +16,9 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
 
+import org.apache.apex.malhar.lib.stream.DevNullCounter;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -25,11 +26,10 @@ import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNullCounter;
 
 /**
  *
- * Functional tests for {@link com.datatorrent.lib.testbench.DevNullCounter}.
+ * Functional tests for {@link org.apache.apex.malhar.lib.testbench.DevNullCounter}.
  * <p>
  * <br>
  * oper.process is called a billion times<br>
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java
index c716206..1fca7e5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/IntegerOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
 
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultOutputPort;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java
index 2e5bcf9..37936e6 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamDuplicaterApp.java
@@ -16,8 +16,10 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.stream.StreamDuplicater;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -25,8 +27,6 @@ import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.stream.StreamDuplicater;
 
 /**
  * Benchmark App for StreamDuplicater Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java
index bb1d081..de9507a 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/stream/StreamMergeApp.java
@@ -16,8 +16,10 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.stream;
+package org.apache.apex.benchmark.stream;
 
+import org.apache.apex.benchmark.WordCountOperator;
+import org.apache.apex.malhar.lib.stream.StreamMerger;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -25,8 +27,6 @@ import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.benchmark.WordCountOperator;
-import com.datatorrent.lib.stream.StreamMerger;
 
 /**
  * Benchmark App for StreamMerge Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java
index b1ddbee..d12267e 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierApp.java
@@ -16,11 +16,13 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.ArrayList;
 import java.util.HashMap;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.EventClassifier;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -28,8 +30,6 @@ import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.EventClassifier;
 
 /**
  * Benchmark App for EventClassifier Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java
index 5fe478b..c1405fc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleApp.java
@@ -16,10 +16,13 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.HashMap;
 
+import org.apache.apex.benchmark.WordCountOperator;
+import org.apache.apex.benchmark.stream.IntegerOperator;
+import org.apache.apex.malhar.lib.testbench.EventClassifierNumberToHashDouble;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -27,9 +30,6 @@ import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.benchmark.WordCountOperator;
-import com.datatorrent.benchmark.stream.IntegerOperator;
-import com.datatorrent.lib.testbench.EventClassifierNumberToHashDouble;
 
 /**
  * Benchmark App for EventClassifierNumberToHashDouble Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java
index 8f28ae6..0c997dc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventGeneratorApp.java
@@ -16,10 +16,12 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.HashMap;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.EventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -27,8 +29,6 @@ import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.EventGenerator;
 
 /**
  * Benchmark App for EventGenerator Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java
index e562224..d7aa2dc 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/EventIncrementerApp.java
@@ -16,19 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.ArrayList;
 import java.util.HashMap;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.EventIncrementer;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.EventIncrementer;
 
 /**
  * Benchmark App for EventIncrementer Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java
index ea2943f..57664ac 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilterClassifierApp.java
@@ -16,19 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.ArrayList;
 import java.util.HashMap;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.FilterClassifier;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.FilterClassifier;
 
 /**
  * Benchmark App for FilterClassifier Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java
index 52c0bed..f282838 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierApp.java
@@ -16,19 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.ArrayList;
 import java.util.HashMap;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.FilteredEventClassifier;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.FilteredEventClassifier;
 
 /**
  * Benchmark App for FilteredEventClassifier Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java
index 29cd079..66fab36 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/HashMapOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.ArrayList;
 import java.util.HashMap;
@@ -24,10 +24,11 @@ import java.util.HashMap;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.testbench.EventGenerator;
+
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.InputOperator;
-import com.datatorrent.lib.testbench.EventGenerator;
 
 /**
  * HashMap Input Operator used as a helper in testbench benchmarking apps.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java
index df5b11e..bdca047 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/RandomEventGeneratorApp.java
@@ -16,16 +16,16 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
 
 /**
  * Benchmark App for RandomEventGenerator Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java
index faafcbf..0b7f7b5 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/SeedEventGeneratorApp.java
@@ -16,10 +16,14 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.ArrayList;
 import java.util.HashMap;
+
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.SeedEventGenerator;
+import org.apache.apex.malhar.lib.util.KeyValPair;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.PortContext;
@@ -27,9 +31,6 @@ import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.SeedEventGenerator;
-import com.datatorrent.lib.util.KeyValPair;
 
 
 /**
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java
index d6e762e..4ad35f9 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/testbench/ThroughputCounterApp.java
@@ -16,18 +16,18 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.util.HashMap;
 
+import org.apache.apex.malhar.lib.stream.DevNull;
+import org.apache.apex.malhar.lib.testbench.ThroughputCounter;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.stream.DevNull;
-import com.datatorrent.lib.testbench.ThroughputCounter;
 
 /**
  * Benchmark App for ThroughputCounter Operator.
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java
index bfdc0ec..a12354e 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractGenerator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
 
 import java.nio.ByteBuffer;
 import java.util.Random;
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java
index 7a345fe..dc0ad48 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/AbstractWindowedOperatorBenchmarkApp.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
 
 import java.io.IOException;
 import java.io.Serializable;
@@ -25,10 +25,13 @@ import org.joda.time.Duration;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.benchmark.window.WindowedOperatorBenchmarkApp.WindowedGenerator;
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
 import org.apache.apex.malhar.lib.state.managed.UnboundedTimeBucketAssigner;
 import org.apache.apex.malhar.lib.state.spillable.SpillableComplexComponentImpl;
 import org.apache.apex.malhar.lib.state.spillable.SpillableStateStore;
 import org.apache.apex.malhar.lib.state.spillable.managed.ManagedTimeUnifiedStateSpillableStateStore;
+import org.apache.apex.malhar.lib.stream.DevNull;
 import org.apache.apex.malhar.lib.window.Accumulation;
 import org.apache.apex.malhar.lib.window.TriggerOption;
 import org.apache.apex.malhar.lib.window.WindowOption;
@@ -47,9 +50,6 @@ import com.datatorrent.api.Operator;
 import com.datatorrent.api.Stats.OperatorStats;
 import com.datatorrent.api.StatsListener;
 import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.benchmark.window.WindowedOperatorBenchmarkApp.WindowedGenerator;
-import com.datatorrent.lib.fileaccess.TFileImpl;
-import com.datatorrent.lib.stream.DevNull;
 
 /**
  * @since 3.7.0
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java
index 7e38584..b8dce7f 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkApp.java
@@ -16,14 +16,16 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
 import org.apache.apex.malhar.lib.state.spillable.SpillableComplexComponentImpl;
 import org.apache.apex.malhar.lib.state.spillable.SpillableSetMultimapImpl;
 import org.apache.apex.malhar.lib.state.spillable.managed.ManagedStateSpillableStateStore;
+import org.apache.apex.malhar.lib.util.KeyValPair;
 import org.apache.apex.malhar.lib.utils.serde.GenericSerde;
 import org.apache.apex.malhar.lib.window.Accumulation;
 import org.apache.apex.malhar.lib.window.Tuple;
@@ -39,8 +41,6 @@ import org.apache.hadoop.conf.Configuration;
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.lib.fileaccess.TFileImpl;
-import com.datatorrent.lib.util.KeyValPair;
 
 /**
  * @since 3.7.0
diff --git a/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java b/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java
index 7e402fc..15a577e 100644
--- a/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java
+++ b/benchmark/src/main/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkApp.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java
index cd8a3ec..644cd01 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/ApplicationFixedTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import java.io.IOException;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java
index 0a21a7c..c5249a2 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/BenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import java.io.IOException;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java
index 6a1c968..aef17b3 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/CouchBaseBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark;
+package org.apache.apex.benchmark;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java
index e2936fe..3417b58 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloApp.java
@@ -16,17 +16,17 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.accumulo;
+package org.apache.apex.benchmark.accumulo;
 
 import org.apache.accumulo.core.data.Mutation;
+import org.apache.apex.malhar.contrib.accumulo.AbstractAccumuloOutputOperator;
+import org.apache.apex.malhar.contrib.accumulo.AccumuloRowTupleGenerator;
+import org.apache.apex.malhar.contrib.accumulo.AccumuloTestHelper;
+import org.apache.apex.malhar.contrib.accumulo.AccumuloTuple;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.contrib.accumulo.AbstractAccumuloOutputOperator;
-import com.datatorrent.contrib.accumulo.AccumuloRowTupleGenerator;
-import com.datatorrent.contrib.accumulo.AccumuloTestHelper;
-import com.datatorrent.contrib.accumulo.AccumuloTuple;
 
 /**
  * BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java
index 8b47a9b..0f48d16 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/accumulo/AccumuloAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.accumulo;
+package org.apache.apex.benchmark.accumulo;
 
 import org.junit.Test;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java
index 14fe441..9eea1b7 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/aerospike/AerospikeBenchmarkAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.aerospike;
+package org.apache.apex.benchmark.aerospike;
 
 import org.junit.Test;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java
index 079d073..56c0cf2 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/algo/UniqueValueCountBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.algo;
+package org.apache.apex.benchmark.algo;
 
 import org.junit.Test;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java
index ec4f308..1708eaf 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/cassandra/CassandraApplicatonTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.cassandra;
+package org.apache.apex.benchmark.cassandra;
 
 import org.junit.Test;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java
index 32a4907..94bc6a9 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseApplicationTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.hbase;
+package org.apache.apex.benchmark.hbase;
 
 import org.junit.Test;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java
index b61f1d3..b6ec667 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hbase/HBaseCsvMappingApplication.java
@@ -16,16 +16,16 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.hbase;
+package org.apache.apex.benchmark.hbase;
 
+import org.apache.apex.malhar.contrib.hbase.HBaseCsvMappingPutOperator;
+import org.apache.apex.malhar.contrib.hbase.HBaseRowStringGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.hbase.HBaseCsvMappingPutOperator;
-import com.datatorrent.contrib.hbase.HBaseRowStringGenerator;
 
 /**
  * BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java
index 653c6f6..aa50e45 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveInsertBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java
index e0097c6..6bd2edf 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/hive/HiveMapBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.hive;
+package org.apache.apex.benchmark.hive;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java
index 6cb901a..8f2c42c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaInputBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java
index 4de7193..1823a00 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/kafka/KafkaOutputBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.kafka;
+package org.apache.apex.benchmark.kafka;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java
index 9201cd5..413438a 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmark.java
@@ -16,18 +16,18 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.contrib.memsql.MemsqlInputOperator;
+import org.apache.apex.malhar.lib.stream.DevNull;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.memsql.MemsqlInputOperator;
-import com.datatorrent.lib.stream.DevNull;
 
 /**
  * BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java
index a596903..f3127a5 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlInputBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
 
 import java.io.FileInputStream;
 import java.io.IOException;
@@ -28,6 +28,9 @@ import org.junit.Test;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.contrib.memsql.AbstractMemsqlOutputOperatorTest;
+import org.apache.apex.malhar.contrib.memsql.MemsqlPOJOOutputOperator;
+import org.apache.apex.malhar.contrib.memsql.MemsqlStore;
 import org.apache.commons.io.IOUtils;
 import org.apache.hadoop.conf.Configuration;
 
@@ -35,15 +38,12 @@ import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.LocalMode;
 import com.datatorrent.api.Operator.ProcessingMode;
-import com.datatorrent.contrib.memsql.AbstractMemsqlOutputOperatorTest;
-import com.datatorrent.contrib.memsql.MemsqlPOJOOutputOperator;
-import com.datatorrent.contrib.memsql.MemsqlStore;
 import com.datatorrent.netlet.util.DTThrowable;
 
-import static com.datatorrent.contrib.memsql.AbstractMemsqlOutputOperatorTest.BATCH_SIZE;
-import static com.datatorrent.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.APP_ID;
-import static com.datatorrent.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.OPERATOR_ID;
-import static com.datatorrent.lib.helper.OperatorContextTestHelper.mockOperatorContext;
+import static org.apache.apex.malhar.contrib.memsql.AbstractMemsqlOutputOperatorTest.BATCH_SIZE;
+import static org.apache.apex.malhar.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.APP_ID;
+import static org.apache.apex.malhar.lib.db.jdbc.JdbcNonTransactionalOutputOperatorTest.OPERATOR_ID;
+import static org.apache.apex.malhar.lib.helper.OperatorContextTestHelper.mockOperatorContext;
 
 public class MemsqlInputBenchmarkTest
 {
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java
index 297bc6d..0b4f126 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmark.java
@@ -16,19 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.contrib.memsql.MemsqlPOJOOutputOperator;
+import org.apache.apex.malhar.lib.testbench.RandomEventGenerator;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
 import com.datatorrent.api.annotation.ApplicationAnnotation;
-import com.datatorrent.contrib.memsql.MemsqlPOJOOutputOperator;
-import com.datatorrent.lib.testbench.RandomEventGenerator;
 
 /**
  * BenchMark Results
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java
index bf82ab3..3d4766e 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/memsql/MemsqlOutputBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.memsql;
+package org.apache.apex.benchmark.memsql;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
@@ -27,12 +27,12 @@ import org.junit.Test;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.contrib.memsql.AbstractMemsqlOutputOperatorTest;
+import org.apache.apex.malhar.contrib.memsql.MemsqlStore;
 import org.apache.commons.io.IOUtils;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.LocalMode;
-import com.datatorrent.contrib.memsql.AbstractMemsqlOutputOperatorTest;
-import com.datatorrent.contrib.memsql.MemsqlStore;
 import com.datatorrent.netlet.util.DTThrowable;
 
 public class MemsqlOutputBenchmarkTest
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java
index d270e7f..904ea6b 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/script/RubyOperatorBenchmarkAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.script;
+package org.apache.apex.benchmark.script;
 
 import org.junit.Test;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java
index cd2c640..6490aa9 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableBenchmarkAppTester.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
 
 import java.io.File;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java
index b87fec1..f6ecf0c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/spillable/SpillableDSBenchmarkTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.spillable;
+package org.apache.apex.benchmark.spillable;
 
 import java.util.Random;
 
@@ -27,14 +27,13 @@ import org.junit.Test;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.fileaccess.TFileImpl;
 import org.apache.apex.malhar.lib.state.spillable.SpillableMapImpl;
 import org.apache.apex.malhar.lib.state.spillable.SpillableTestUtils;
 import org.apache.apex.malhar.lib.state.spillable.managed.ManagedStateSpillableStateStore;
 import org.apache.apex.malhar.lib.utils.serde.Serde;
 import org.apache.apex.malhar.lib.utils.serde.StringSerde;
 
-import com.datatorrent.lib.fileaccess.TFileImpl;
-
 public class SpillableDSBenchmarkTest
 {
   private static final Logger logger = LoggerFactory.getLogger(SpillableDSBenchmarkTest.class);
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java
index dc8f4b4..eb3cc7a 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/state/ManagedStateBenchmarkAppTest.java
@@ -16,20 +16,20 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.state;
+package org.apache.apex.benchmark.state;
 
 import java.io.File;
 
 import org.junit.Before;
 import org.junit.Test;
 
+import org.apache.apex.benchmark.state.StoreOperator.ExecMode;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileUtil;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.LocalMode;
 import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.benchmark.state.StoreOperator.ExecMode;
 
 /**
  * This is not a really unit test, but in fact a benchmark runner.
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java
index 99d8a1f..88d66b4 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java
index 929d8bc..9fb4638 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventClassifierNumberToHashDoubleAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java
index 5a427a5..0a71ea4 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventGeneratorAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java
index 1a85a7b..fb4401e 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/EventIncrementerAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java
index 9419022..c921a4c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilterClassifierAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java
index 977d6b7..658e372 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/FilteredEventClassifierAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java
index 92ca0fd..b278f6d 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/testbench/ThroughputCounterAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.testbench;
+package org.apache.apex.benchmark.testbench;
 
 import java.io.FileInputStream;
 import java.io.FileNotFoundException;
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java
index 157accc..64fca9e 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/util/serde/GenericSerdePerformanceTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.util.serde;
+package org.apache.apex.benchmark.util.serde;
 
 import java.util.Random;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java
index 2bc9335..fe9ba0c 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/window/KeyedWindowedOperatorBenchmarkAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
 
 import java.io.File;
 
diff --git a/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java b/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java
index 4a16396..b6f3549 100644
--- a/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java
+++ b/benchmark/src/test/java/org/apache/apex/benchmark/window/WindowedOperatorBenchmarkAppTest.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.benchmark.window;
+package org.apache.apex.benchmark.window;
 
 import java.io.File;
 
diff --git a/contrib/pom.xml b/contrib/pom.xml
index a69ad29..13c77b9 100755
--- a/contrib/pom.xml
+++ b/contrib/pom.xml
@@ -25,7 +25,7 @@
   <parent>
     <groupId>org.apache.apex</groupId>
     <artifactId>malhar</artifactId>
-    <version>3.9.0-SNAPSHOT</version>
+    <version>4.0.0-SNAPSHOT</version>
   </parent>
 
   <artifactId>malhar-contrib</artifactId>
@@ -36,7 +36,6 @@
     <!-- skip tests by default as they depend on manual setup -->
     <skip.contrib.module.tests>false</skip.contrib.module.tests>
     <skipTests>true</skipTests>
-    <checkstyle.console>false</checkstyle.console>
   </properties>
   <repositories>
     <repository>
@@ -217,14 +216,6 @@
         <artifactId>japicmp-maven-plugin</artifactId>
       </plugin>
 
-      <plugin>
-        <groupId>org.apache.maven.plugins</groupId>
-        <artifactId>maven-checkstyle-plugin</artifactId>
-        <configuration>
-          <maxAllowedViolations>2709</maxAllowedViolations>
-          <logViolationsToConsole>${checkstyle.console}</logViolationsToConsole>
-        </configuration>
-      </plugin>
     </plugins>
   </build>
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java
index 50a1056..8250cd0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
 
 import java.util.Map.Entry;
 
@@ -24,8 +24,8 @@ import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.Scanner;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Value;
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
 
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
 /**
  * Base input adapter, which reads data from persistence database and writes into output port(s).&nbsp; Subclasses should provide the
  * implementation of getting the tuples and scanner methods.
@@ -38,18 +38,21 @@ import com.datatorrent.lib.db.AbstractStoreInputOperator;
  * @tags key value, accumulo
  * @since 1.0.4
  */
-public abstract class AbstractAccumuloInputOperator<T> extends AbstractStoreInputOperator<T, AccumuloStore> {
+public abstract class AbstractAccumuloInputOperator<T> extends AbstractStoreInputOperator<T, AccumuloStore>
+{
 
   public abstract T getTuple(Entry<Key, Value> entry);
 
   public abstract Scanner getScanner(Connector conn);
 
-  public AbstractAccumuloInputOperator() {
+  public AbstractAccumuloInputOperator()
+  {
     store = new AccumuloStore();
   }
 
   @Override
-  public void emitTuples() {
+  public void emitTuples()
+  {
     Connector conn = store.getConnector();
     Scanner scan = getScanner(conn);
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java
index a4ae1ac..153f88d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AbstractAccumuloOutputOperator.java
@@ -16,19 +16,21 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
 
 import java.util.List;
 
-import org.apache.accumulo.core.client.MutationsRejectedException;
-import org.apache.accumulo.core.data.Mutation;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.accumulo.core.client.MutationsRejectedException;
+import org.apache.accumulo.core.data.Mutation;
+import org.apache.apex.malhar.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
+
+import com.google.common.collect.Lists;
+
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
-import com.google.common.collect.Lists;
 
 /**
  * Base output operator that stores tuples in Accumulo rows.&nbsp; Subclasses should provide implementation of operationMutation method. <br>
@@ -56,7 +58,8 @@ import com.google.common.collect.Lists;
  *            The tuple type
  * @since 1.0.4
  */
-public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregateTransactionableStoreOutputOperator<T, AccumuloWindowStore> {
+public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregateTransactionableStoreOutputOperator<T, AccumuloWindowStore>
+{
   private static final transient Logger logger = LoggerFactory.getLogger(AbstractAccumuloOutputOperator.class);
   private final List<T> tuples;
   private transient ProcessingMode mode;
@@ -75,6 +78,7 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
     tuples = Lists.newArrayList();
     store = new AccumuloWindowStore();
   }
+
   @Override
   public void processTuple(T tuple)
   {
@@ -82,7 +86,8 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
   }
 
   @Override
-  public void storeAggregate() {
+  public void storeAggregate()
+  {
     try {
       for (T tuple : tuples) {
         Mutation mutation = operationMutation(tuple);
@@ -96,6 +101,7 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
     }
     tuples.clear();
   }
+
   /**
    *
    * @param t
@@ -106,11 +112,11 @@ public abstract class AbstractAccumuloOutputOperator<T> extends AbstractAggregat
   @Override
   public void setup(OperatorContext context)
   {
-    mode=context.getValue(context.PROCESSING_MODE);
-    if(mode==ProcessingMode.EXACTLY_ONCE){
+    mode = context.getValue(context.PROCESSING_MODE);
+    if (mode == ProcessingMode.EXACTLY_ONCE) {
       throw new RuntimeException("This operator only supports atmost once and atleast once processing modes");
     }
-    if(mode==ProcessingMode.AT_MOST_ONCE){
+    if (mode == ProcessingMode.AT_MOST_ONCE) {
       tuples.clear();
     }
     super.setup(context);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java
index 52d273a..2eb4db0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloStore.java
@@ -16,10 +16,13 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
 
 import java.io.IOException;
 
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.BatchWriter;
@@ -30,11 +33,9 @@ import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.ZooKeeperInstance;
 import org.apache.accumulo.core.client.security.tokens.PasswordToken;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.Connectable;
 
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.Connectable;
 
 /**
  * A {@link Connectable} for accumulo that implements Connectable interface.
@@ -45,7 +46,8 @@ import com.datatorrent.lib.db.Connectable;
  * @param <T>
  * @since 1.0.4
  */
-public class AccumuloStore implements Connectable {
+public class AccumuloStore implements Connectable
+{
   private static final transient Logger logger = LoggerFactory.getLogger(AccumuloStore.class);
   private String zookeeperHost;
   private String instanceName;
@@ -58,12 +60,13 @@ public class AccumuloStore implements Connectable {
 
   private long memoryLimit;
   private int numThreads;
-  private static final long DEFAULT_MEMORY=2147483648l;
-  private static final int DEFAULT_THREADS=1;
+  private static final long DEFAULT_MEMORY = 2147483648L;
+  private static final int DEFAULT_THREADS = 1;
 
-  public AccumuloStore(){
-    memoryLimit=DEFAULT_MEMORY;
-    numThreads=DEFAULT_THREADS;
+  public AccumuloStore()
+  {
+    memoryLimit = DEFAULT_MEMORY;
+    numThreads = DEFAULT_THREADS;
   }
 
   /**
@@ -71,7 +74,8 @@ public class AccumuloStore implements Connectable {
    *
    * @return Connector
    */
-  public Connector getConnector() {
+  public Connector getConnector()
+  {
     return connector;
   }
 
@@ -80,7 +84,8 @@ public class AccumuloStore implements Connectable {
    *
    * @return TableName
    */
-  public String getTableName() {
+  public String getTableName()
+  {
     return tableName;
   }
 
@@ -89,7 +94,8 @@ public class AccumuloStore implements Connectable {
    *
    * @param tableName
    */
-  public void setTableName(String tableName) {
+  public void setTableName(String tableName)
+  {
     this.tableName = tableName;
   }
 
@@ -98,7 +104,8 @@ public class AccumuloStore implements Connectable {
    *
    * @return ZookeeperHost
    */
-  public String getZookeeperHost() {
+  public String getZookeeperHost()
+  {
     return zookeeperHost;
   }
 
@@ -107,7 +114,8 @@ public class AccumuloStore implements Connectable {
    *
    * @param zookeeperHost
    */
-  public void setZookeeperHost(String zookeeperHost) {
+  public void setZookeeperHost(String zookeeperHost)
+  {
     this.zookeeperHost = zookeeperHost;
   }
 
@@ -116,7 +124,8 @@ public class AccumuloStore implements Connectable {
    *
    * @return instanceName
    */
-  public String getInstanceName() {
+  public String getInstanceName()
+  {
     return instanceName;
   }
 
@@ -125,7 +134,8 @@ public class AccumuloStore implements Connectable {
    *
    * @param instanceName
    */
-  public void setInstanceName(String instanceName) {
+  public void setInstanceName(String instanceName)
+  {
     this.instanceName = instanceName;
   }
 
@@ -134,7 +144,8 @@ public class AccumuloStore implements Connectable {
    *
    * @param userName
    */
-  public void setUserName(String userName) {
+  public void setUserName(String userName)
+  {
     this.userName = userName;
   }
 
@@ -143,23 +154,28 @@ public class AccumuloStore implements Connectable {
    *
    * @param password
    */
-  public void setPassword(String password) {
+  public void setPassword(String password)
+  {
     this.password = password;
   }
+
   /**
    * setter for memory limit
    *
    * @param memoryLimit
    */
-  public void setMemoryLimit(long memoryLimit) {
+  public void setMemoryLimit(long memoryLimit)
+  {
     this.memoryLimit = memoryLimit;
   }
+
   /**
    * setter for number of writer threads
    *
    * @param numThreads
    */
-  public void setNumThreads(int numThreads) {
+  public void setNumThreads(int numThreads)
+  {
     this.numThreads = numThreads;
   }
 
@@ -168,12 +184,14 @@ public class AccumuloStore implements Connectable {
    *
    * @return BatchWriter
    */
-  public BatchWriter getBatchwriter() {
+  public BatchWriter getBatchwriter()
+  {
     return batchwriter;
   }
 
   @Override
-  public void connect() throws IOException {
+  public void connect() throws IOException
+  {
     Instance instance = null;
     instance = new ZooKeeperInstance(instanceName, zookeeperHost);
     try {
@@ -198,7 +216,8 @@ public class AccumuloStore implements Connectable {
   }
 
   @Override
-  public void disconnect() throws IOException {
+  public void disconnect() throws IOException
+  {
     try {
       batchwriter.close();
     } catch (MutationsRejectedException e) {
@@ -208,7 +227,8 @@ public class AccumuloStore implements Connectable {
   }
 
   @Override
-  public boolean isConnected() {
+  public boolean isConnected()
+  {
     // Not applicable for accumulo
     return false;
   }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java
index 9bec031..412ae14 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/AccumuloWindowStore.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
 
 import java.io.ByteArrayInputStream;
 import java.io.ByteArrayOutputStream;
@@ -25,6 +25,9 @@ import java.io.DataOutputStream;
 import java.io.IOException;
 import java.util.Map.Entry;
 
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.Scanner;
 import org.apache.accumulo.core.client.TableNotFoundException;
@@ -33,12 +36,10 @@ import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Range;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.security.Authorizations;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
 import org.apache.hadoop.io.Text;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.TransactionableStore;
 
 /**
  * Provides transactional support by implementing TransactionableStore abstract methods.
@@ -53,7 +54,8 @@ import com.datatorrent.lib.db.TransactionableStore;
  * @tags accumulo, key value
  * @since 1.0.4
  */
-public class AccumuloWindowStore extends AccumuloStore implements TransactionableStore {
+public class AccumuloWindowStore extends AccumuloStore implements TransactionableStore
+{
   private static final transient Logger logger = LoggerFactory.getLogger(AccumuloWindowStore.class);
   private static final String DEFAULT_ROW_NAME = "AccumuloOperator_row";
   private static final String DEFAULT_COLUMN_FAMILY_NAME = "AccumuloOutputOperator_cf";
@@ -67,7 +69,8 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
   private transient String lastWindowColumnName;
   private transient byte[] lastWindowColumnBytes;
 
-  public AccumuloWindowStore() {
+  public AccumuloWindowStore()
+  {
     rowName = DEFAULT_ROW_NAME;
     columnFamilyName = DEFAULT_COLUMN_FAMILY_NAME;
     lastWindowColumnName = DEFAULT_LAST_WINDOW_PREFIX_COLUMN_NAME;
@@ -78,54 +81,64 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
    * the values are stored as byte arrays.This method converts string to byte
    * arrays. uses util class in hbase library to do so.
    */
-  private void constructKeys() {
+  private void constructKeys()
+  {
     rowBytes = rowName.getBytes();
     columnFamilyBytes = columnFamilyName.getBytes();
   }
 
-  public String getRowName() {
+  public String getRowName()
+  {
     return rowName;
   }
 
-  public void setRowName(String rowName) {
+  public void setRowName(String rowName)
+  {
     this.rowName = rowName;
     constructKeys();
   }
 
-  public String getColumnFamilyName() {
+  public String getColumnFamilyName()
+  {
     return columnFamilyName;
   }
 
-  public void setColumnFamilyName(String columnFamilyName) {
+  public void setColumnFamilyName(String columnFamilyName)
+  {
     this.columnFamilyName = columnFamilyName;
     constructKeys();
   }
 
   @Override
-  public void beginTransaction() {
+  public void beginTransaction()
+  {
     // accumulo does not support transactions
   }
 
   @Override
-  public void commitTransaction() {
+  public void commitTransaction()
+  {
     // accumulo does not support transactions
 
   }
 
   @Override
-  public void rollbackTransaction() {
+  public void rollbackTransaction()
+  {
     // accumulo does not support transactions
 
   }
 
   @Override
-  public boolean isInTransaction() {
+  public boolean isInTransaction()
+  {
     // accumulo does not support transactions
     return false;
   }
 
   @Override
-  public long getCommittedWindowId(String appId, int operatorId) {
+  public long getCommittedWindowId(String appId, int operatorId)
+  {
     byte[] value = null;
     Authorizations auths = new Authorizations();
     Scanner scan = null;
@@ -150,7 +163,8 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
   }
 
   @Override
-  public void storeCommittedWindowId(String appId, int operatorId,long windowId) {
+  public void storeCommittedWindowId(String appId, int operatorId,long windowId)
+  {
     byte[] WindowIdBytes = toBytes(windowId);
     String columnKey = appId + "_" + operatorId + "_" + lastWindowColumnName;
     lastWindowColumnBytes = columnKey.getBytes();
@@ -166,19 +180,19 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
   }
 
   @Override
-  public void removeCommittedWindowId(String appId, int operatorId) {
+  public void removeCommittedWindowId(String appId, int operatorId)
+  {
     // accumulo does not support transactions
-
   }
 
   public static byte[] toBytes(long l)
   {
-    ByteArrayOutputStream baos=new ByteArrayOutputStream(Long.SIZE/8);
-    DataOutputStream dos=new DataOutputStream(baos);
-    byte[] result=null;
+    ByteArrayOutputStream baos = new ByteArrayOutputStream(Long.SIZE / 8);
+    DataOutputStream dos = new DataOutputStream(baos);
+    byte[] result = null;
     try {
       dos.writeLong(l);
-      result=baos.toByteArray();
+      result = baos.toByteArray();
       dos.close();
     } catch (IOException e) {
       logger.error("error converting to byte array");
@@ -187,11 +201,11 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
     return result;
   }
 
-
-  public static long toLong(byte[] b){
-    ByteArrayInputStream baos=new ByteArrayInputStream(b);
-    DataInputStream dos=new DataInputStream(baos);
-    long result=0;
+  public static long toLong(byte[] b)
+  {
+    ByteArrayInputStream baos = new ByteArrayInputStream(b);
+    DataInputStream dos = new DataInputStream(baos);
+    long result = 0;
     try {
       result = dos.readLong();
       dos.close();
@@ -202,5 +216,4 @@ public class AccumuloWindowStore extends AccumuloStore implements Transactionabl
     return result;
   }
 
-
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java
index cbac406..508c16e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/accumulo/package-info.java
@@ -17,4 +17,4 @@
  * under the License.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.accumulo;
+package org.apache.apex.malhar.contrib.accumulo;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java
index 9f7469f..3cc0d69 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeGetOperator.java
@@ -16,31 +16,34 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
 
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
 
 import com.aerospike.client.Record;
 import com.aerospike.client.query.RecordSet;
 import com.aerospike.client.query.Statement;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
 /**
- * Base input adapter, which reads data from persistence database through its API and writes into output port(s).&nsbsp;
- * Subclasses should provide the implementation of getting the tuples and querying to retrieve data.
+ * Base input adapter, which reads data from persistence database through its
+ * API and writes into output port(s). Subclasses should provide the
+ * implementation of getting the tuples and querying to retrieve data.
  * <p>
- * This is an abstract class. Sub-classes need to implement {@link #queryToRetrieveData()} and {@link #getTuple(Record)}.
- * </p>
+ * This is an abstract class. Sub-classes need to implement
+ * {@link #queryToRetrieveData()} and {@link #getTuple(Record)}.
+ *
  * @displayName Abstract Aerospike Get
  * @category Input
  * @tags get
  * @since 1.0.4
  */
-public abstract class AbstractAerospikeGetOperator<T> extends AbstractStoreInputOperator<T, AerospikeStore> {
-
+public abstract class AbstractAerospikeGetOperator<T> extends AbstractStoreInputOperator<T, AerospikeStore>
+{
   private static final Logger logger = LoggerFactory.getLogger(AbstractAerospikeGetOperator.class);
 
   /**
@@ -69,20 +72,19 @@ public abstract class AbstractAerospikeGetOperator<T> extends AbstractStoreInput
    * It then converts each row into tuple and emit that into output port.
    */
   @Override
-  public void emitTuples() {
-
+  public void emitTuples()
+  {
     Statement query = queryToRetrieveData();
     logger.debug(String.format("select statement: %s", query.toString()));
     RecordSet rs;
     try {
       rs = store.getClient().query(null, query);
-      while(rs.next()){
+      while (rs.next()) {
         Record rec = rs.getRecord();
         T tuple = getTuple(rec);
         outputPort.emit(tuple);
       }
-    }
-    catch (Exception ex) {
+    } catch (Exception ex) {
       store.disconnect();
       DTThrowable.rethrow(ex);
     }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java
index a629078..ca14a37 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeNonTransactionalPutOperator.java
@@ -16,32 +16,36 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
 
 import java.util.List;
 
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
+
 import com.aerospike.client.AerospikeException;
 import com.aerospike.client.Bin;
 import com.aerospike.client.Key;
 import com.google.common.collect.Lists;
 
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
-
 
 /**
- * <p>
- * Generic base output adaptor which writes tuples as they come without providing any transactional support.&nbsp; Subclasses should provide implementation for getting updated bins.
- * </p>
+ * Generic base output adaptor which writes tuples as they come without
+ * providing any transactional support. Subclasses should provide implementation
+ * for getting updated bins.
+ *
  * @displayName Abstract Aerospike Non Transactional Put
  * @category Output
  * @tags put, non transactional
  * @param <T> type of tuple
  * @since 1.0.4
  */
-public abstract class AbstractAerospikeNonTransactionalPutOperator<T> extends AbstractStoreOutputOperator<T,AerospikeStore> {
+public abstract class AbstractAerospikeNonTransactionalPutOperator<T>
+    extends AbstractStoreOutputOperator<T,AerospikeStore>
+{
+  private final transient List<Bin> bins;
 
-  private transient final List<Bin> bins;
-  public AbstractAerospikeNonTransactionalPutOperator() {
+  public AbstractAerospikeNonTransactionalPutOperator()
+  {
     super();
     bins = Lists.newArrayList();
   }
@@ -58,8 +62,8 @@ public abstract class AbstractAerospikeNonTransactionalPutOperator<T> extends Ab
   protected abstract Key getUpdatedBins(T tuple, List<Bin> bins) throws AerospikeException;
 
   @Override
-  public void processTuple(T tuple) {
-
+  public void processTuple(T tuple)
+  {
     Key key;
     Bin[] binsArray;
     try {
@@ -68,11 +72,9 @@ public abstract class AbstractAerospikeNonTransactionalPutOperator<T> extends Ab
       binsArray = bins.toArray(binsArray);
       store.getClient().put(null, key, binsArray);
       bins.clear();
-    }
-    catch (AerospikeException e) {
+    } catch (AerospikeException e) {
       throw new RuntimeException(e);
     }
-
   }
 
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java
index 6db8896..9b8ba59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AbstractAerospikeTransactionalPutOperator.java
@@ -16,39 +16,42 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
 
 import java.util.Collection;
 import java.util.List;
 
+import org.apache.apex.malhar.lib.db.AbstractBatchTransactionableStoreOutputOperator;
+
 import com.aerospike.client.AerospikeException;
 import com.aerospike.client.Bin;
 import com.aerospike.client.Key;
 import com.google.common.collect.Lists;
 
-import com.datatorrent.lib.db.AbstractBatchTransactionableStoreOutputOperator;
-
 /**
+ * Generic base adaptor which creates a transaction at the start of window.
+ * Subclasses should provide implementation for getting updated bins.
  * <p>
- * Generic base adaptor which creates a transaction at the start of window.&nbsp; Subclasses should provide implementation for getting updated bins. <br/>
- * </p>
- * <p>
- * Executes all the put updates and closes the transaction at the end of the window.
- * The tuples in a window are stored in check-pointed collection which is cleared in the endWindow().
- * This is needed for the recovery. The operator writes a tuple at least once in the database, which is why
- * only when all the updates are executed, the transaction is committed in the end window call.
+ * Executes all the put updates and closes the transaction at the end of the
+ * window. The tuples in a window are stored in check-pointed collection which
+ * is cleared in the endWindow(). This is needed for the recovery. The operator
+ * writes a tuple at least once in the database, which is why only when all the
+ * updates are executed, the transaction is committed in the end window call.
  * </p>
+ *
  * @displayName Abstract Aerospike Transactional Put
  * @category Output
  * @tags put, transactional
  * @param <T>type of tuple
  * @since 1.0.4
  */
-public abstract class AbstractAerospikeTransactionalPutOperator<T> extends AbstractBatchTransactionableStoreOutputOperator<T, AerospikeTransactionalStore> {
-
-  private transient final List<Bin> bins;
+public abstract class AbstractAerospikeTransactionalPutOperator<T>
+    extends AbstractBatchTransactionableStoreOutputOperator<T, AerospikeTransactionalStore>
+{
+  private final transient List<Bin> bins;
 
-  public AbstractAerospikeTransactionalPutOperator() {
+  public AbstractAerospikeTransactionalPutOperator()
+  {
     super();
     bins = Lists.newArrayList();
   }
@@ -70,7 +73,7 @@ public abstract class AbstractAerospikeTransactionalPutOperator<T> extends Abstr
     Key key;
     Bin[] binsArray;
     try {
-      for(T tuple: tuples) {
+      for (T tuple: tuples) {
         key = getUpdatedBins(tuple,bins);
         binsArray = new Bin[bins.size()];
         binsArray = bins.toArray(binsArray);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java
index f69e433..49580d0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJONonTransactionalPutOperator.java
@@ -16,22 +16,22 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
 
 import java.util.ArrayList;
 import java.util.List;
 
 import javax.validation.constraints.NotNull;
 
-import com.aerospike.client.Bin;
-import com.aerospike.client.Key;
-
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
+
+import com.aerospike.client.Bin;
+import com.aerospike.client.Key;
 
 /**
  * <p>
@@ -47,8 +47,8 @@ import com.datatorrent.lib.util.PojoUtils.Getter;
 @Evolving
 public class AerospikePOJONonTransactionalPutOperator extends AbstractAerospikeNonTransactionalPutOperator<Object>
 {
-  private static transient final Logger LOG
-    = LoggerFactory.getLogger(AerospikePOJONonTransactionalPutOperator.class);
+  private static final Logger LOG =
+      LoggerFactory.getLogger(AerospikePOJONonTransactionalPutOperator.class);
 
   // Two element list; first retrieves the record key and second the list of bins in this tuple
   @NotNull
@@ -91,7 +91,7 @@ public class AerospikePOJONonTransactionalPutOperator extends AbstractAerospikeN
     }
     Key key = keyGetter.get(tuple);
     List<Bin> binList = binsGetter.get(tuple);
-    if ( ! (null == binList || binList.isEmpty()) ) {
+    if (!(null == binList || binList.isEmpty())) {
       list.addAll(binList);
     }
     return key;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java
index f59b893..7add642 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikePOJOTransactionalPutOperator.java
@@ -16,27 +16,24 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
 
 import java.util.ArrayList;
 import java.util.List;
 
 import javax.validation.constraints.NotNull;
 
-import com.aerospike.client.Bin;
-import com.aerospike.client.Key;
-
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
 import org.apache.hadoop.classification.InterfaceStability.Evolving;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
+import com.aerospike.client.Bin;
+import com.aerospike.client.Key;
 
 /**
  * <p>
  * A generic implementation of
- * {@link com.datatorrent.contrib.aerospike.AbstractAerospikeTransactionalPutOperator} which can
+ * {@link org.apache.apex.malhar.contrib.aerospike.AbstractAerospikeTransactionalPutOperator} which can
  * take a POJO.
  * </p>
  * @displayName Aerospike Transactional Put Operator
@@ -45,12 +42,8 @@ import com.datatorrent.lib.util.PojoUtils.Getter;
  * @since 2.1.0
  */
 @Evolving
-public class AerospikePOJOTransactionalPutOperator
-  extends AbstractAerospikeTransactionalPutOperator<Object>
+public class AerospikePOJOTransactionalPutOperator extends AbstractAerospikeTransactionalPutOperator<Object>
 {
-  private static transient final Logger LOG
-    = LoggerFactory.getLogger(AerospikePOJOTransactionalPutOperator.class);
-
   // Two element list; first retrieves the record key and second the list of bins in this tuple
   @NotNull
   private ArrayList<String> expressions;
@@ -92,11 +85,10 @@ public class AerospikePOJOTransactionalPutOperator
     }
     Key key = keyGetter.get(tuple);
     List<Bin> binList = binsGetter.get(tuple);
-    if ( ! (null == binList || binList.isEmpty()) ) {
+    if ( !(null == binList || binList.isEmpty()) ) {
       list.addAll(binList);
     }
     return key;
   }
 
-
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java
index 10809e1..2ffda24 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeStore.java
@@ -16,15 +16,18 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
 
 import javax.validation.constraints.NotNull;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.Connectable;
+
 import com.aerospike.client.AerospikeClient;
 import com.aerospike.client.AerospikeException;
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.Connectable;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
 /**
  * A {@link Connectable} that uses aerospike to connect to stores and implements Connectable interface.
@@ -34,8 +37,8 @@ import org.slf4j.LoggerFactory;
  * @tags store
  * @since 1.0.4
  */
-public class AerospikeStore implements Connectable {
-
+public class AerospikeStore implements Connectable
+{
   protected static final Logger logger = LoggerFactory.getLogger(AerospikeStore.class);
   private String userName;
   private String password;
@@ -49,7 +52,8 @@ public class AerospikeStore implements Connectable {
    *
    * @param userName user name.
    */
-  public void setUserName(String userName) {
+  public void setUserName(String userName)
+  {
     this.userName = userName;
   }
 
@@ -58,7 +62,8 @@ public class AerospikeStore implements Connectable {
    *
    * @param password password
    */
-  public void setPassword(String password) {
+  public void setPassword(String password)
+  {
     this.password = password;
   }
 
@@ -68,7 +73,8 @@ public class AerospikeStore implements Connectable {
    * @return The node
    */
   @NotNull
-  public String getNode() {
+  public String getNode()
+  {
     return node;
   }
 
@@ -77,7 +83,8 @@ public class AerospikeStore implements Connectable {
    *
    * @param node node
    */
-  public void setNode(@NotNull String node) {
+  public void setNode(@NotNull String node)
+  {
     this.node = node;
   }
 
@@ -86,7 +93,8 @@ public class AerospikeStore implements Connectable {
    *
    * @return The client
    */
-  public AerospikeClient getClient() {
+  public AerospikeClient getClient()
+  {
     return client;
   }
 
@@ -95,7 +103,8 @@ public class AerospikeStore implements Connectable {
    *
    * @param port port
    */
-  public void setPort(int port) {
+  public void setPort(int port)
+  {
     this.port = port;
   }
 
@@ -103,15 +112,14 @@ public class AerospikeStore implements Connectable {
    * Create connection with database.
    */
   @Override
-  public void connect() {
+  public void connect()
+  {
     try {
       client = new AerospikeClient(node, port);
       logger.debug("Aerospike connection Success");
-    }
-    catch (AerospikeException ex) {
+    } catch (AerospikeException ex) {
       throw new RuntimeException("closing database resource", ex);
-    }
-    catch (Throwable t) {
+    } catch (Throwable t) {
       DTThrowable.rethrow(t);
     }
   }
@@ -120,12 +128,14 @@ public class AerospikeStore implements Connectable {
    * Close connection.
    */
   @Override
-  public void disconnect() {
+  public void disconnect()
+  {
     client.close();
   }
 
   @Override
-  public boolean isConnected() {
+  public boolean isConnected()
+  {
     return !client.isConnected();
   }
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java
index 8128a17..8f898e6 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/AerospikeTransactionalStore.java
@@ -16,10 +16,12 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
 
 import javax.annotation.Nonnull;
 
+import org.apache.apex.malhar.lib.db.TransactionableStore;
+
 import com.aerospike.client.AerospikeException;
 import com.aerospike.client.Bin;
 import com.aerospike.client.Key;
@@ -28,7 +30,6 @@ import com.aerospike.client.query.IndexType;
 import com.aerospike.client.query.RecordSet;
 import com.aerospike.client.query.Statement;
 import com.aerospike.client.task.IndexTask;
-import com.datatorrent.lib.db.TransactionableStore;
 
 /**
  * <p>Provides transaction support to the operators by implementing TransactionableStore abstract methods. </p>
@@ -38,8 +39,8 @@ import com.datatorrent.lib.db.TransactionableStore;
  * @tags store, transactional
  * @since 1.0.4
  */
-public class AerospikeTransactionalStore extends AerospikeStore implements TransactionableStore {
-
+public class AerospikeTransactionalStore extends AerospikeStore implements TransactionableStore
+{
   public static String DEFAULT_APP_ID_COL = "dt_app_id";
   public static String DEFAULT_OPERATOR_ID_COL = "dt_operator_id";
   public static String DEFAULT_WINDOW_COL = "dt_window";
@@ -59,8 +60,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
   private transient boolean inTransaction;
   private transient Statement lastWindowFetchCommand;
 
-  public AerospikeTransactionalStore() {
-
+  public AerospikeTransactionalStore()
+  {
     super();
     metaSet = DEFAULT_META_SET;
     metaTableAppIdColumn = DEFAULT_APP_ID_COL;
@@ -75,8 +76,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
    *
    * @param metaSet meta set name.
    */
-  public void setMetaSet(@Nonnull String metaSet) {
-
+  public void setMetaSet(@Nonnull String metaSet)
+  {
     this.metaSet = metaSet;
   }
 
@@ -86,8 +87,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
    *
    * @param appIdColumn application id column name.
    */
-  public void setMetaTableAppIdColumn(@Nonnull String appIdColumn) {
-
+  public void setMetaTableAppIdColumn(@Nonnull String appIdColumn)
+  {
     this.metaTableAppIdColumn = appIdColumn;
   }
 
@@ -97,8 +98,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
    *
    * @param operatorIdColumn operator id column name.
    */
-  public void setMetaTableOperatorIdColumn(@Nonnull String operatorIdColumn) {
-
+  public void setMetaTableOperatorIdColumn(@Nonnull String operatorIdColumn)
+  {
     this.metaTableOperatorIdColumn = operatorIdColumn;
   }
 
@@ -108,8 +109,8 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
    *
    * @param windowColumn window column name.
    */
-  public void setMetaTableWindowColumn(@Nonnull String windowColumn) {
-
+  public void setMetaTableWindowColumn(@Nonnull String windowColumn)
+  {
     this.metaTableWindowColumn = windowColumn;
   }
 
@@ -118,14 +119,14 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
    *
    * @param namespace namespace.
    */
-  public void setNamespace(@Nonnull String namespace) {
-
+  public void setNamespace(@Nonnull String namespace)
+  {
     this.namespace = namespace;
   }
 
   @Override
-  public void connect() {
-
+  public void connect()
+  {
     super.connect();
     createIndexes();
     try {
@@ -133,44 +134,43 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
       lastWindowFetchCommand.setNamespace(namespace);
       lastWindowFetchCommand.setSetName(metaSet);
       lastWindowFetchCommand.setBinNames(metaTableWindowColumn);
-    }
-    catch (Exception e) {
+    } catch (Exception e) {
       throw new RuntimeException(e);
     }
   }
 
   @Override
-  public void disconnect() {
-
+  public void disconnect()
+  {
     super.disconnect();
   }
 
   @Override
-  public void beginTransaction() {
-
+  public void beginTransaction()
+  {
     inTransaction = true;
   }
 
   @Override
-  public void commitTransaction() {
-
+  public void commitTransaction()
+  {
     inTransaction = false;
   }
 
   @Override
-  public void rollbackTransaction() {
-
+  public void rollbackTransaction()
+  {
     inTransaction = false;
   }
 
   @Override
-  public boolean isInTransaction() {
-
+  public boolean isInTransaction()
+  {
     return inTransaction;
   }
 
-  private void createIndexes() {
-
+  private void createIndexes()
+  {
     IndexTask task;
     try {
       task = client.createIndex(null, namespace, metaSet,
@@ -186,26 +186,25 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
   }
 
   @Override
-  public long getCommittedWindowId(String appId, int operatorId) {
-
+  public long getCommittedWindowId(String appId, int operatorId)
+  {
     try {
       lastWindowFetchCommand.setFilters(Filter.equal(metaTableOperatorIdColumn, operatorId));
       lastWindowFetchCommand.setFilters(Filter.equal(metaTableAppIdColumn, appId));
       long lastWindow = -1;
       RecordSet recordSet = client.query(null, lastWindowFetchCommand);
-      while(recordSet.next()) {
+      while (recordSet.next()) {
         lastWindow = Long.parseLong(recordSet.getRecord().getValue(metaTableWindowColumn).toString());
       }
       return lastWindow;
-    }
-    catch (AerospikeException ex) {
+    } catch (AerospikeException ex) {
       throw new RuntimeException(ex);
     }
   }
 
   @Override
-  public void storeCommittedWindowId(String appId, int operatorId, long windowId) {
-
+  public void storeCommittedWindowId(String appId, int operatorId, long windowId)
+  {
     try {
       String keyString = appId + String.valueOf(operatorId);
       Key key = new Key(namespace,metaSet,keyString.hashCode());
@@ -213,21 +212,19 @@ public class AerospikeTransactionalStore extends AerospikeStore implements Trans
       Bin bin2 = new Bin(metaTableOperatorIdColumn,operatorId);
       Bin bin3 = new Bin(metaTableWindowColumn,windowId);
       client.put(null, key, bin1,bin2,bin3);
-    }
-    catch (AerospikeException e) {
+    } catch (AerospikeException e) {
       throw new RuntimeException(e);
     }
   }
 
   @Override
-  public void removeCommittedWindowId(String appId, int operatorId) {
-
+  public void removeCommittedWindowId(String appId, int operatorId)
+  {
     try {
       String keyString = appId + String.valueOf(operatorId);
       Key key = new Key(namespace,metaSet,keyString.hashCode());
       client.delete(null, key);
-    }
-    catch (AerospikeException e) {
+    } catch (AerospikeException e) {
       throw new RuntimeException(e);
     }
   }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java
index 97e878c..8a37041 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/aerospike/package-info.java
@@ -17,4 +17,4 @@
  * under the License.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.aerospike;
+package org.apache.apex.malhar.contrib.aerospike;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java
index ecd57aa..7f44cc9 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApacheLogInputGenerator.java
@@ -16,18 +16,30 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.apachelog;
-
-import com.datatorrent.api.*;
-import com.datatorrent.api.Context.OperatorContext;
-import java.io.*;
+package org.apache.apex.malhar.contrib.apachelog;
+
+import java.io.BufferedReader;
+import java.io.File;
+import java.io.FileInputStream;
+import java.io.IOException;
+import java.io.InputStream;
+import java.io.InputStreamReader;
 import java.text.SimpleDateFormat;
-import java.util.*;
+import java.util.ArrayList;
+import java.util.Date;
+import java.util.List;
+import java.util.Random;
+import java.util.StringTokenizer;
 import java.util.concurrent.ArrayBlockingQueue;
 import javax.validation.constraints.NotNull;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import com.datatorrent.api.Context.OperatorContext;
+import com.datatorrent.api.DefaultOutputPort;
+import com.datatorrent.api.InputOperator;
+import com.datatorrent.api.Operator;
+
 /**
  * An implementation of input operator and activation listener that simulates the apache logs.
  * <p>
@@ -38,7 +50,7 @@ import org.slf4j.LoggerFactory;
  */
 public class ApacheLogInputGenerator implements InputOperator, Operator.ActivationListener<OperatorContext>
 {
-  private final static String delimiter = ";";
+  private static final String delimiter = ";";
 
   private transient Random random;
   private transient int ipAddressCount;
@@ -99,13 +111,12 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
 
   private List<String> readLines(String file) throws IOException
   {
-    List<String> lines = new ArrayList<String>();
+    List<String> lines = new ArrayList<>();
     InputStream in;
     File f = new File(file);
     if (f.exists()) {
       in = new FileInputStream(f);
-    }
-    else {
+    } else {
       in = getClass().getResourceAsStream(file);
     }
     BufferedReader br = new BufferedReader(new InputStreamReader(in));
@@ -114,8 +125,7 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
       while ((line = br.readLine()) != null) {
         lines.add(line);
       }
-    }
-    finally {
+    } finally {
       br.close();
     }
     return lines;
@@ -150,8 +160,7 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
         bytes.add(Integer.parseInt(token.nextToken().trim()));
         status.add(Integer.parseInt(token.nextToken().trim()));
       }
-    }
-    catch (IOException e) {
+    } catch (IOException e) {
       throw new RuntimeException(e);
     }
 
@@ -214,8 +223,7 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
           if (maxDelay > 0) {
             try {
               Thread.sleep(random.nextInt(maxDelay));
-            }
-            catch (InterruptedException e) {
+            } catch (InterruptedException e) {
               return;
             }
           }
@@ -232,8 +240,8 @@ public class ApacheLogInputGenerator implements InputOperator, Operator.Activati
     try {
       thread.interrupt();
       thread.join();
-    }
-    catch (InterruptedException ex) {
+    } catch (InterruptedException ex) {
+      // ignore
     }
   }
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java
index 23b4a62..b8d29d4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLocalLog.java
@@ -16,15 +16,16 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
+
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.io.fs.TailFsInputOperator;
+import org.apache.apex.malhar.lib.logs.ApacheLogParseMapOutputOperator;
+import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
 import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.io.fs.TailFsInputOperator;
-import com.datatorrent.lib.logs.ApacheLogParseMapOutputOperator;
-import org.apache.hadoop.conf.Configuration;
 
 /**
  * An implementation of Streaming Application that reads from a local apache log file on the fly and dumps the parsed data to output.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java
index d624e08..0eef6dc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/ApplicationLogGenerator.java
@@ -16,16 +16,16 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
 
-import com.datatorrent.api.*;
-import com.datatorrent.api.DAG.Locality;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.logs.ApacheLogParseMapOutputOperator;
-import java.util.ArrayList;
-import java.util.List;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.logs.ApacheLogParseMapOutputOperator;
 import org.apache.hadoop.conf.Configuration;
 
+import com.datatorrent.api.DAG;
+import com.datatorrent.api.DAG.Locality;
+import com.datatorrent.api.StreamingApplication;
+
 /**
  * An implementation of Streaming Application that generates apache log file on the fly and dumps the parsed data to output.
  *
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java
index bd226e3..a58c39e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/GeoIPExtractor.java
@@ -16,19 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
 
+import java.io.IOException;
 import java.util.HashMap;
 import java.util.Map;
-import java.io.IOException;
 import javax.validation.constraints.NotNull;
 
-import com.datatorrent.lib.logs.InformationExtractor;
-import com.maxmind.geoip.Location;
-import com.maxmind.geoip.LookupService;
-
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.logs.InformationExtractor;
+
+import com.maxmind.geoip.Location;
+import com.maxmind.geoip.LookupService;
 
 /**
  * An implementation of InformationExtractor that extracts Geo information from an IP address using maxmind API .
@@ -67,8 +67,7 @@ public class GeoIPExtractor implements InformationExtractor
   {
     try {
       reader = new LookupService(databasePath, LookupService.GEOIP_MEMORY_CACHE | LookupService.GEOIP_CHECK_CACHE);
-    }
-    catch (IOException ex) {
+    } catch (IOException ex) {
       throw new RuntimeException(ex);
     }
   }
@@ -90,8 +89,7 @@ public class GeoIPExtractor implements InformationExtractor
         m.put("ipRegion", location.region);
         m.put("ipCity", location.city);
       }
-    }
-    catch (Exception ex) {
+    } catch (Exception ex) {
       LOG.error("Caught exception when looking up Geo IP for {}:", value, ex);
     }
     return m;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java
index bb87f03..1b6b4b6 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/TimestampExtractor.java
@@ -16,20 +16,20 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
 
-import java.util.Map;
-
-import com.datatorrent.lib.logs.InformationExtractor;
 import java.text.DateFormat;
 import java.text.ParseException;
 import java.text.SimpleDateFormat;
 import java.util.Date;
 import java.util.HashMap;
+import java.util.Map;
 import javax.validation.constraints.NotNull;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.logs.InformationExtractor;
+
 /**
  * An implementation of InformationExtractor that extracts the time stamp in milliseconds from epoch from an arbitrary date string.
  * <p>
@@ -69,8 +69,7 @@ public class TimestampExtractor implements InformationExtractor
     try {
       Date date = dateFormat.parse((String)value);
       m.put("timestamp", date.getTime());
-    }
-    catch (ParseException ex) {
+    } catch (ParseException ex) {
       LOG.error("Error parsing \"{}\" to timestamp using \"{}\":", value, dateFormatString, ex);
     }
     return m;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java
index f73ff3d..799eb48 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/UserAgentExtractor.java
@@ -16,20 +16,20 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
 
 import java.util.HashMap;
 import java.util.Map;
+import java.util.concurrent.TimeUnit;
+
+import org.apache.apex.malhar.lib.logs.InformationExtractor;
 
-import com.datatorrent.lib.logs.InformationExtractor;
 import com.google.common.cache.Cache;
 import com.google.common.cache.CacheBuilder;
 
-import java.util.concurrent.TimeUnit;
-
 import net.sf.uadetector.ReadableUserAgent;
-import net.sf.uadetector.service.UADetectorServiceFactory;
 import net.sf.uadetector.UserAgentStringParser;
+import net.sf.uadetector.service.UADetectorServiceFactory;
 
 /**
  * An implementation of Information Extractor that extracts the browser and the OS from a user-agent string.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java
index d99c7b5..67b151b 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/package-info.java
@@ -20,4 +20,4 @@
  * Apache log application, operators, and utilities.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.apachelog;
+package org.apache.apex.malhar.contrib.apachelog;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java
index 738a146..e44c97a 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/apachelog/zmq/Application.java
@@ -16,21 +16,21 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.apachelog.zmq;
+package org.apache.apex.malhar.contrib.apachelog.zmq;
 
 
+import org.apache.apex.malhar.contrib.zmq.SimpleSinglePortZeroMQPullStringInputOperator;
+import org.apache.apex.malhar.lib.algo.UniqueCounter;
+import org.apache.apex.malhar.lib.io.ConsoleOutputOperator;
+import org.apache.apex.malhar.lib.logs.ApacheLogParseOperator;
+import org.apache.apex.malhar.lib.math.Sum;
 import org.apache.hadoop.conf.Configuration;
 
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DAG;
 import com.datatorrent.api.DAG.Locality;
-import com.datatorrent.api.annotation.ApplicationAnnotation;
 import com.datatorrent.api.StreamingApplication;
-import com.datatorrent.contrib.zmq.SimpleSinglePortZeroMQPullStringInputOperator;
-import com.datatorrent.lib.algo.UniqueCounter;
-import com.datatorrent.lib.io.ConsoleOutputOperator;
-import com.datatorrent.lib.logs.ApacheLogParseOperator;
-import com.datatorrent.lib.math.Sum;
+import com.datatorrent.api.annotation.ApplicationAnnotation;
 
 /**
  * <p>An implementation of Streaming Application that populates the DAG. </p>
@@ -40,7 +40,7 @@ import com.datatorrent.lib.math.Sum;
  * @tag: streaming
  * @since 0.3.2
  */
-@ApplicationAnnotation(name="ApacheLog")
+@ApplicationAnnotation(name = "ApacheLog")
 public class Application implements StreamingApplication
 {
   private Locality locality = null;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java
index f863d41..52200fe 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
 
 import java.io.IOException;
 import java.io.InputStream;
@@ -24,6 +24,7 @@ import java.io.InputStream;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.io.fs.AbstractFileInputOperator;
 import org.apache.apex.malhar.lib.wal.FSWindowDataManager;
 import org.apache.avro.AvroRuntimeException;
 import org.apache.avro.file.DataFileStream;
@@ -37,7 +38,6 @@ import com.google.common.annotations.VisibleForTesting;
 
 import com.datatorrent.api.AutoMetric;
 import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.lib.io.fs.AbstractFileInputOperator;
 
 /**
  * <p>
diff --git a/contrib/src/main/java/com/datatorrent/contrib/avro/AvroFileToPojoModule.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileToPojoModule.java
similarity index 98%
rename from contrib/src/main/java/com/datatorrent/contrib/avro/AvroFileToPojoModule.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileToPojoModule.java
index 8ad00df..4637b8f 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/avro/AvroFileToPojoModule.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroFileToPojoModule.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
 
 import org.apache.apex.malhar.lib.wal.FSWindowDataManager;
 import org.apache.avro.generic.GenericRecord;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java
index 012e977..176f5eb 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroRecordHelper.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
 
 import java.text.ParseException;
 import java.util.List;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java
index 2acf98c..5539562 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/AvroToPojo.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
 
 import java.util.ArrayList;
 import java.util.List;
@@ -25,6 +25,9 @@ import java.util.StringTokenizer;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.FieldInfo.SupportType;
+import org.apache.apex.malhar.lib.util.PojoUtils;
 import org.apache.avro.AvroRuntimeException;
 import org.apache.avro.generic.GenericRecord;
 import org.apache.commons.lang3.ClassUtils;
@@ -40,9 +43,6 @@ import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
 import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.FieldInfo.SupportType;
-import com.datatorrent.lib.util.PojoUtils;
 
 /**
  * <p>
@@ -87,7 +87,7 @@ public class AvroToPojo extends BaseOperator
   @AutoMetric
   @VisibleForTesting
   int fieldErrorCount = 0;
-  
+
   public final transient DefaultOutputPort<GenericRecord> errorPort = new DefaultOutputPort<GenericRecord>();
 
   /**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java
index 41c56e3..17a4ecc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/avro/PojoToAvro.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.avro;
+package org.apache.apex.malhar.contrib.avro;
 
 import java.io.IOException;
 import java.util.ArrayList;
@@ -25,6 +25,8 @@ import java.util.List;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
 import org.apache.avro.AvroRuntimeException;
 import org.apache.avro.Schema;
 import org.apache.avro.Schema.Field;
@@ -42,8 +44,6 @@ import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.annotation.InputPortFieldAnnotation;
 import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
 
 /**
  * <p>
@@ -63,7 +63,6 @@ import com.datatorrent.lib.util.PojoUtils.Getter;
 @InterfaceStability.Evolving
 public class PojoToAvro extends BaseOperator
 {
-
   private List<Field> columnNames;
 
   private Class<?> cls;
@@ -155,7 +154,7 @@ public class PojoToAvro extends BaseOperator
    * @return Getter
    */
   private Getter<?, ?> generateGettersForField(Class<?> cls, String inputFieldName)
-      throws NoSuchFieldException, SecurityException
+    throws NoSuchFieldException, SecurityException
   {
     java.lang.reflect.Field f = cls.getDeclaredField(inputFieldName);
     Class<?> c = ClassUtils.primitiveToWrapper(f.getType());
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java
index 7bd47fc..3153387 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraInputOperator.java
@@ -16,18 +16,18 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
 
 import com.datastax.driver.core.PagingState;
 import com.datastax.driver.core.ResultSet;
 import com.datastax.driver.core.Row;
 import com.datastax.driver.core.SimpleStatement;
 
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.netlet.util.DTThrowable;
 
@@ -43,8 +43,8 @@ import com.datatorrent.netlet.util.DTThrowable;
  * @tags cassandra
  * @since 1.0.2
  */
-public abstract class AbstractCassandraInputOperator<T> extends AbstractStoreInputOperator<T, CassandraStore> {
-
+public abstract class AbstractCassandraInputOperator<T> extends AbstractStoreInputOperator<T, CassandraStore>
+{
   private static final Logger logger = LoggerFactory.getLogger(AbstractCassandraInputOperator.class);
   private PagingState nextPageState;
   private int fetchSize;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java
index 9048383..f0f590d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractCassandraTransactionableOutputOperator.java
@@ -16,12 +16,14 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import java.util.Collection;
 
 import javax.annotation.Nonnull;
 
+import org.apache.apex.malhar.lib.db.AbstractBatchTransactionableStoreOutputOperator;
+
 import com.datastax.driver.core.BatchStatement;
 import com.datastax.driver.core.PreparedStatement;
 import com.datastax.driver.core.Statement;
@@ -29,7 +31,6 @@ import com.datastax.driver.core.exceptions.DriverException;
 import com.datatorrent.api.Context;
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.Operator.ActivationListener;
-import com.datatorrent.lib.db.AbstractBatchTransactionableStoreOutputOperator;
 
 /**
  * <p>
@@ -86,8 +87,7 @@ public abstract class AbstractCassandraTransactionableOutputOperator<T> extends
   public void processBatch(Collection<T> tuples)
   {
     BatchStatement batchCommand = store.getBatchCommand();
-    for(T tuple: tuples)
-    {
+    for (T tuple: tuples) {
       batchCommand.add(setStatementParameters(updateCommand, tuple));
     }
   }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java
index f287385..cadf157 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/AbstractUpsertOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import java.io.IOException;
 import java.io.InputStream;
@@ -949,7 +949,7 @@ public abstract class AbstractUpsertOutputOperator extends BaseOperator implemen
 
   /**
    * Implementing concrete Operator instances define the Connection Builder properties by implementing this method
-   * Please refer to {@link com.datatorrent.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} for
+   * Please refer to {@link org.apache.apex.malhar.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} for
    * an example implementation of the ConnectionStateManager instantiation.
    * Note that if this method is returning null, the connection properties are
    * fetched from a properties file loaded from the classpath.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java
index f43777f..cab83e8 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOInputOperator.java
@@ -16,17 +16,30 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import java.math.BigDecimal;
-import java.util.*;
+import java.util.ArrayList;
+import java.util.Date;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.UUID;
 
 import javax.validation.constraints.Min;
 import javax.validation.constraints.NotNull;
 
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Setter;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterBoolean;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterDouble;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterFloat;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterInt;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterLong;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
 
 import com.datastax.driver.core.BoundStatement;
 import com.datastax.driver.core.ColumnDefinitions;
@@ -39,9 +52,6 @@ import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.Operator;
 import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.*;
 
 /**
  * <p>
@@ -287,8 +297,7 @@ public class CassandraPOJOInputOperator extends AbstractCassandraInputOperator<O
     try {
       // This code will be replaced after integration of creating POJOs on the fly utility.
       obj = pojoClass.newInstance();
-    }
-    catch (InstantiationException | IllegalAccessException ex) {
+    } catch (InstantiationException | IllegalAccessException ex) {
       throw new RuntimeException(ex);
     }
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java
index a191bb0..5f7d52a 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPOJOOutputOperator.java
@@ -16,28 +16,44 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import java.lang.reflect.Field;
 import java.math.BigDecimal;
-import java.util.*;
+import java.util.ArrayList;
+import java.util.Date;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.UUID;
 
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.google.common.collect.Lists;
-import com.datastax.driver.core.*;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterBoolean;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterDouble;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterFloat;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterInt;
+import org.apache.apex.malhar.lib.util.PojoUtils.GetterLong;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
+
+import com.datastax.driver.core.BoundStatement;
+import com.datastax.driver.core.ColumnDefinitions;
+import com.datastax.driver.core.DataType;
+import com.datastax.driver.core.LocalDate;
+import com.datastax.driver.core.PreparedStatement;
+import com.datastax.driver.core.Statement;
 import com.datastax.driver.core.exceptions.DriverException;
+import com.google.common.collect.Lists;
 import com.datatorrent.api.AutoMetric;
 import com.datatorrent.api.Context;
 import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.annotation.InputPortFieldAnnotation;
 import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.*;
 
 /**
  * <p>
@@ -106,10 +122,11 @@ public class CassandraPOJOOutputOperator extends AbstractCassandraTransactionabl
   @Override
   public void activate(Context.OperatorContext context)
   {
-    com.datastax.driver.core.ResultSet rs = store.getSession().execute("select * from " + store.keyspace + "." + tablename);
+    com.datastax.driver.core.ResultSet rs
+        = store.getSession().execute("select * from " + store.keyspace + "." + tablename);
     final ColumnDefinitions rsMetaData = rs.getColumnDefinitions();
 
-    if(fieldInfos == null) {
+    if (fieldInfos == null) {
       populateFieldInfosFromPojo(rsMetaData);
     }
 
@@ -228,13 +245,12 @@ public class CassandraPOJOOutputOperator extends AbstractCassandraTransactionabl
       }
     }
     String statement
-            = "INSERT INTO " + store.keyspace + "."
-            + tablename
-            + " (" + queryfields.toString() + ") "
-            + "VALUES (" + values.toString() + ");";
+        = "INSERT INTO " + store.keyspace + "."
+        + tablename
+        + " (" + queryfields.toString() + ") "
+        + "VALUES (" + values.toString() + ");";
     LOG.debug("statement is {}", statement);
     return store.getSession().prepare(statement);
-
   }
 
   @Override
@@ -316,6 +332,7 @@ public class CassandraPOJOOutputOperator extends AbstractCassandraTransactionabl
       errorRecords++;
     }
   }
+
   /**
    * A list of {@link FieldInfo}s where each item maps a column name to a pojo field name.
    */
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java
index 8b435c9..179ac1a 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPojoUtils.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import java.math.BigDecimal;
 import java.util.Date;
@@ -25,11 +25,12 @@ import java.util.Map;
 import java.util.Set;
 import java.util.UUID;
 
+import org.apache.apex.malhar.lib.util.PojoUtils;
+
 import com.datastax.driver.core.BoundStatement;
 import com.datastax.driver.core.DataType;
 import com.datastax.driver.core.LocalDate;
 import com.datastax.driver.core.TypeCodec;
-import com.datatorrent.lib.util.PojoUtils;
 
 /**
  * Used to manage simple data type based getters for given cassandra columns
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java
index 713179c..a797bab 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraPreparedStatementGenerator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 
 import java.util.EnumSet;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java
index 5d9178c..0c298aa 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraStore.java
@@ -16,19 +16,20 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import javax.validation.constraints.NotNull;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.db.Connectable;
+
 import com.datastax.driver.core.Cluster;
 import com.datastax.driver.core.ProtocolVersion;
 import com.datastax.driver.core.Session;
 import com.datastax.driver.core.exceptions.DriverException;
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.Connectable;
 
 /**
  * A {@link Connectable} that uses cassandra to connect to stores and implements Connectable interface.
@@ -67,7 +68,8 @@ public class CassandraStore implements Connectable
    *
    * @param keyspace keyspace.
    */
-  public void setKeyspace(String keyspace) {
+  public void setKeyspace(String keyspace)
+  {
     this.keyspace = keyspace;
   }
 
@@ -106,7 +108,8 @@ public class CassandraStore implements Connectable
   }
 
   @NotNull
-  public String getNode() {
+  public String getNode()
+  {
     return node;
   }
 
@@ -115,15 +118,18 @@ public class CassandraStore implements Connectable
    *
    * @param node node
    */
-  public void setNode(@NotNull String node) {
+  public void setNode(@NotNull String node)
+  {
     this.node = node;
   }
 
-  public Cluster getCluster() {
+  public Cluster getCluster()
+  {
     return cluster;
   }
 
-  public Session getSession() {
+  public Session getSession()
+  {
     return session;
   }
 
@@ -167,20 +173,18 @@ public class CassandraStore implements Connectable
   public void connect()
   {
     try {
-      if(cluster==null)
+      if (cluster == null) {
         buildCluster();
+      }
       session = cluster.connect();
       logger.debug("Cassandra connection Success");
-    }
-    catch (DriverException ex) {
+    } catch (DriverException ex) {
       throw new RuntimeException("closing database resource", ex);
-    }
-    catch (Throwable t) {
+    } catch (Throwable t) {
       DTThrowable.rethrow(t);
     }
   }
 
-
   /**
    * Close connection.
    */
@@ -190,11 +194,9 @@ public class CassandraStore implements Connectable
     try {
       session.close();
       cluster.close();
-    }
-    catch (DriverException ex) {
+    } catch (DriverException ex) {
       throw new RuntimeException("closing database resource", ex);
-    }
-    catch (Throwable t) {
+    } catch (Throwable t) {
       DTThrowable.rethrow(t);
     }
   }
@@ -204,8 +206,7 @@ public class CassandraStore implements Connectable
   {
     try {
       return !session.isClosed();
-    }
-    catch (DriverException ex) {
+    } catch (DriverException ex) {
       throw new RuntimeException("closing database resource", ex);
     }
   }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java
index 4c5a71e..29b6ce9 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/CassandraTransactionalStore.java
@@ -16,14 +16,18 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import javax.annotation.Nonnull;
 
-import com.datastax.driver.core.*;
-import com.datastax.driver.core.exceptions.DriverException;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
 
-import com.datatorrent.lib.db.TransactionableStore;
+import com.datastax.driver.core.BatchStatement;
+import com.datastax.driver.core.BoundStatement;
+import com.datastax.driver.core.PreparedStatement;
+import com.datastax.driver.core.ResultSet;
+import com.datastax.driver.core.Statement;
+import com.datastax.driver.core.exceptions.DriverException;
 
 /**
  * <p>Provides transaction support to the operators by implementing TransactionableStore abstract methods. </p>
@@ -34,8 +38,8 @@ import com.datatorrent.lib.db.TransactionableStore;
  * @tags cassandra, transactional
  * @since 1.0.2
  */
-public class CassandraTransactionalStore extends CassandraStore implements TransactionableStore {
-
+public class CassandraTransactionalStore extends CassandraStore implements TransactionableStore
+{
   public static String DEFAULT_APP_ID_COL = "dt_app_id";
   public static String DEFAULT_OPERATOR_ID_COL = "dt_operator_id";
   public static String DEFAULT_WINDOW_COL = "dt_window";
@@ -116,7 +120,8 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
     this.metaTableWindowColumn = windowColumn;
   }
 
-  public Statement getLastWindowUpdateStatement() {
+  public Statement getLastWindowUpdateStatement()
+  {
     return lastWindowUpdateStatement;
   }
 
@@ -130,26 +135,24 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
   {
     super.connect();
     try {
-      String command = "SELECT " + metaTableWindowColumn + " FROM " + keyspace +"."+ metaTable + " WHERE " + metaTableAppIdColumn +
+      String command = "SELECT " + metaTableWindowColumn + " FROM " + keyspace + "." + metaTable +
+          " WHERE " + metaTableAppIdColumn +
           " = ? AND " + metaTableOperatorIdColumn + " = ?";
       logger.debug(command);
       lastWindowFetchCommand = session.prepare(command);
 
-      command = "UPDATE " + keyspace +"."+ metaTable + " SET " + metaTableWindowColumn + " = ? where " + metaTableAppIdColumn + " = ? " +
+      command = "UPDATE " + keyspace + "." + metaTable + " SET " + metaTableWindowColumn + " = ? where " + metaTableAppIdColumn + " = ? " +
           " and " + metaTableOperatorIdColumn + " = ?";
       logger.debug(command);
       lastWindowUpdateCommand = session.prepare(command);
 
-      command = "DELETE FROM " + keyspace +"."+ metaTable + " where " + metaTableAppIdColumn + " = ? and " +
+      command = "DELETE FROM " + keyspace + "." + metaTable + " where " + metaTableAppIdColumn + " = ? and " +
           metaTableOperatorIdColumn + " = ?";
       logger.debug(command);
       lastWindowDeleteCommand = session.prepare(command);
-    }
-    catch (DriverException e) {
+    } catch (DriverException e) {
       throw new RuntimeException(e);
-    }
-    catch (Exception e)
-    {
+    } catch (Exception e) {
       throw new RuntimeException(e);
     }
   }
@@ -160,8 +163,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
     if (lastWindowUpdateCommand != null) {
       try {
         lastWindowUpdateCommand.disableTracing();
-      }
-      catch (DriverException e) {
+      } catch (DriverException e) {
         throw new RuntimeException(e);
       }
     }
@@ -208,8 +210,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
       }
       lastWindowFetchCommand.disableTracing();
       return lastWindow;
-    }
-    catch (DriverException ex) {
+    } catch (DriverException ex) {
       throw new RuntimeException(ex);
     }
   }
@@ -221,8 +222,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
       BoundStatement boundStatement = new BoundStatement(lastWindowUpdateCommand);
       lastWindowUpdateStatement = boundStatement.bind(windowId,appId,operatorId);
       batchCommand.add(lastWindowUpdateStatement);
-    }
-    catch (DriverException e) {
+    } catch (DriverException e) {
       throw new RuntimeException(e);
     }
   }
@@ -235,8 +235,7 @@ public class CassandraTransactionalStore extends CassandraStore implements Trans
       lastWindowDeleteStatement = boundStatement.bind(appId,operatorId);
       session.execute(lastWindowDeleteStatement);
       lastWindowDeleteCommand.disableTracing();
-    }
-    catch (DriverException e) {
+    } catch (DriverException e) {
       throw new RuntimeException(e);
     }
   }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java
index f089137..0eccb22 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/ConnectionStateManager.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import java.io.Serializable;
 import java.net.InetAddress;
@@ -64,7 +64,7 @@ import static com.google.common.base.Preconditions.checkNotNull;
  * </p>
  * <p> A typical implementation of the ConnectionBuilder would like this:
  * <pre>
- *     
+ *
  *     public ConnectionStateManager.ConnectionBuilder withConnectionBuilder()
  *     {
  *       return ConnectionStateManager.withNewBuilder()
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java
index 92be546..10ffb08 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/UpsertExecutionContext.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
 
 import com.datastax.driver.core.ConsistencyLevel;
 
@@ -122,7 +122,7 @@ public class UpsertExecutionContext<T>
 
   /**
    * This decides if we want to override the default TTL if at all set in the
-   * {@link com.datatorrent.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} that is used to execute a
+   * {@link org.apache.apex.malhar.contrib.cassandra.ConnectionStateManager.ConnectionBuilder} that is used to execute a
    * mutation. Note that TTLs are not mandatory for mutations.
    * Also it is supported to have TTLs only for the current execution context but not set a default at the
    * connection state manager level
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java
index 1baae5e..93c5518 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/cassandra/package-info.java
@@ -17,4 +17,4 @@
  * under the License.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.cassandra;
+package org.apache.apex.malhar.contrib.cassandra;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java
index 1cd4eb5..3c53a49 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
 import java.io.IOException;
 import java.util.Collection;
@@ -24,21 +24,20 @@ import java.util.List;
 import java.util.Map;
 import java.util.concurrent.TimeUnit;
 
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
+import org.apache.apex.malhar.lib.util.KryoCloneUtils;
+
 import com.couchbase.client.CouchbaseClient;
 import com.couchbase.client.vbucket.config.Config;
 import com.google.common.annotations.VisibleForTesting;
 import com.google.common.collect.Lists;
 
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
-
 import com.datatorrent.api.Context;
 import com.datatorrent.api.DefaultPartition;
 import com.datatorrent.api.Partitioner;
-
-import com.datatorrent.lib.util.KryoCloneUtils;
 import com.datatorrent.netlet.util.DTThrowable;
 
 /**
@@ -94,8 +93,7 @@ public abstract class AbstractCouchBaseInputOperator<T> extends AbstractStoreInp
       }
       try {
         clientPartition = store.connectServer(serverURIString);
-      }
-      catch (IOException ex) {
+      } catch (IOException ex) {
         DTThrowable.rethrow(ex);
       }
     }
@@ -116,16 +114,16 @@ public abstract class AbstractCouchBaseInputOperator<T> extends AbstractStoreInp
     List<String> keys = getKeys();
     Object result = null;
     for (String key: keys) {
-        int master = conf.getMaster(conf.getVbucketByKey(key));
-        if (master == getServerIndex()) {
-          result = clientPartition.get(key);
-        }
+      int master = conf.getMaster(conf.getVbucketByKey(key));
+      if (master == getServerIndex()) {
+        result = clientPartition.get(key);
       }
+    }
 
-      if (result != null) {
-        T tuple = getTuple(result);
-        outputPort.emit(tuple);
-      }
+    if (result != null) {
+      T tuple = getTuple(result);
+      outputPort.emit(tuple);
+    }
   }
 
   public abstract T getTuple(Object object);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java
index ef3be01..8cc92f2 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseOutputOperator.java
@@ -16,23 +16,24 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
 import java.util.HashMap;
 import java.util.Iterator;
 import java.util.TreeMap;
-import net.spy.memcached.internal.OperationCompletionListener;
-import net.spy.memcached.internal.OperationFuture;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.datatorrent.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
+import org.apache.apex.malhar.lib.db.AbstractAggregateTransactionableStoreOutputOperator;
 
 import com.datatorrent.api.Context.OperatorContext;
-
 import com.datatorrent.netlet.util.DTThrowable;
 
+import net.spy.memcached.internal.OperationCompletionListener;
+import net.spy.memcached.internal.OperationFuture;
+
+
 /**
  * AbstractCouchBaseOutputOperator which extends Transactionable Store Output Operator.
  * Classes extending from this operator should implement the abstract functionality of generateKey, getValue and insertOrUpdate.
@@ -108,10 +109,10 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
     id++;
     String key = getKey(tuple);
     Object value = getValue(tuple);
-    if(!(value instanceof Boolean) && !(value instanceof Integer) && !(value instanceof String) && !(value instanceof Float) && !(value instanceof Double) && !(value instanceof Character) && !(value instanceof Long) && !(value instanceof Short) && !(value instanceof Byte)){
-    if (serializer != null) {
-      value = serializer.serialize(value);
-    }
+    if (!(value instanceof Boolean) && !(value instanceof Integer) && !(value instanceof String) && !(value instanceof Float) && !(value instanceof Double) && !(value instanceof Character) && !(value instanceof Long) && !(value instanceof Short) && !(value instanceof Byte)) {
+      if (serializer != null) {
+        value = serializer.serialize(value);
+      }
     }
     OperationFuture<Boolean> future = processKeyValue(key, value);
     synchronized (syncObj) {
@@ -135,7 +136,7 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
   public void waitForQueueSize(int sizeOfQueue)
   {
     long startTms = System.currentTimeMillis();
-    long elapsedTime ;
+    long elapsedTime;
     while (numTuples > sizeOfQueue) {
       synchronized (syncObj) {
         if (numTuples > sizeOfQueue) {
@@ -143,12 +144,10 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
             elapsedTime = System.currentTimeMillis() - startTms;
             if (elapsedTime >= store.timeout) {
               throw new RuntimeException("Timed out waiting for space in queue");
-            }
-            else {
+            } else {
               syncObj.wait(store.timeout - elapsedTime);
             }
-          }
-          catch (InterruptedException ex) {
+          } catch (InterruptedException ex) {
             DTThrowable.rethrow(ex);
           }
         }
@@ -167,7 +166,7 @@ public abstract class AbstractCouchBaseOutputOperator<T> extends AbstractAggrega
     public void onComplete(OperationFuture<?> f) throws Exception
     {
       if (!((Boolean)f.get())) {
-        logger.error("Operation failed {}" , f);
+        logger.error("Operation failed {}", f);
         failure = true;
         return;
       }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java
index f33d804..0abe06f 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/AbstractCouchBaseSetOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
 import net.spy.memcached.internal.OperationFuture;
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java
index 188d989..0f968bd 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseJSONSerializer.java
@@ -16,11 +16,11 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
-import com.datatorrent.netlet.util.DTThrowable;
 import java.io.IOException;
 import org.codehaus.jackson.map.ObjectMapper;
+import com.datatorrent.netlet.util.DTThrowable;
 
 /**
  * <p>CouchBaseJSONSerializer</p>
@@ -43,8 +43,7 @@ public class CouchBaseJSONSerializer implements CouchBaseSerializer
     String value = null;
     try {
       value = mapper.writeValueAsString(o);
-    }
-    catch (IOException ex) {
+    } catch (IOException ex) {
       DTThrowable.rethrow(ex);
     }
     return value;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java
index 2378aec..91e3b17 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBasePOJOInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
 import java.io.IOException;
 import java.util.Iterator;
@@ -24,12 +24,16 @@ import java.util.Iterator;
 import javax.validation.constraints.Min;
 import javax.validation.constraints.NotNull;
 
-import com.couchbase.client.protocol.views.*;
-
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.codehaus.jackson.map.ObjectMapper;
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
+
+import com.couchbase.client.protocol.views.Query;
+import com.couchbase.client.protocol.views.Stale;
+import com.couchbase.client.protocol.views.View;
+import com.couchbase.client.protocol.views.ViewResponse;
+import com.couchbase.client.protocol.views.ViewRow;
 
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
 import com.datatorrent.api.Context.OperatorContext;
 
 /**
@@ -167,8 +171,7 @@ public class CouchBasePOJOInputOperator extends AbstractStoreInputOperator<Objec
     super.setup(context);
     try {
       className = Class.forName(outputClass);
-    }
-    catch (ClassNotFoundException ex) {
+    } catch (ClassNotFoundException ex) {
       throw new RuntimeException(ex);
     }
     view = store.getInstance().getView(designDocumentName, viewName);
@@ -198,8 +201,7 @@ public class CouchBasePOJOInputOperator extends AbstractStoreInputOperator<Objec
       Object outputObj = null;
       try {
         outputObj = objectMapper.readValue(document.toString(), className);
-      }
-      catch (IOException ex) {
+      } catch (IOException ex) {
         throw new RuntimeException(ex);
       }
       outputPort.emit(outputObj);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java
index f15758e..0769e82 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseSerializer.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
 /**
  *
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java
index 238047e..fb5ba1d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseStore.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
 import java.io.IOException;
 import java.net.URI;
@@ -28,6 +28,10 @@ import java.util.concurrent.TimeUnit;
 import javax.annotation.Nonnull;
 import javax.validation.constraints.Min;
 
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.db.Connectable;
+
 import com.couchbase.client.CouchbaseClient;
 import com.couchbase.client.CouchbaseConnectionFactoryBuilder;
 import com.couchbase.client.vbucket.ConfigurationProvider;
@@ -35,11 +39,6 @@ import com.couchbase.client.vbucket.ConfigurationProviderHTTP;
 import com.couchbase.client.vbucket.config.Bucket;
 import com.couchbase.client.vbucket.config.Config;
 
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import com.datatorrent.lib.db.Connectable;
-
 import com.datatorrent.netlet.util.DTThrowable;
 
 /**
@@ -108,6 +107,7 @@ public class CouchBaseStore implements Connectable
   {
     this.queueSize = queueSize;
   }
+
   protected Integer maxTuples = 1000;
   protected int blockTime = 1000;
   protected long timeout = 10000;
@@ -192,8 +192,7 @@ public class CouchBaseStore implements Connectable
   {
     try {
       connect();
-    }
-    catch (IOException ex) {
+    } catch (IOException ex) {
       DTThrowable.rethrow(ex);
     }
     ConfigurationProvider configurationProvider = new ConfigurationProviderHTTP(baseURIs, userConfig, passwordConfig);
@@ -201,8 +200,7 @@ public class CouchBaseStore implements Connectable
     Config conf = configBucket.getConfig();
     try {
       disconnect();
-    }
-    catch (IOException ex) {
+    } catch (IOException ex) {
       DTThrowable.rethrow(ex);
     }
     return conf;
@@ -216,8 +214,7 @@ public class CouchBaseStore implements Connectable
     for (String url : tokens) {
       try {
         uri = new URI("http", url, "/pools", null, null);
-      }
-      catch (URISyntaxException ex) {
+      } catch (URISyntaxException ex) {
         DTThrowable.rethrow(ex);
       }
       baseURIs.add(uri);
@@ -228,9 +225,8 @@ public class CouchBaseStore implements Connectable
       cfb.setOpQueueMaxBlockTime(blockTime); // wait up to 10 second when trying to enqueue an operation
       client = new CouchbaseClient(cfb.buildCouchbaseConnection(baseURIs, bucket, password));
       //client = new CouchbaseClient(baseURIs, "default", "");
-    }
-    catch (IOException e) {
-      logger.error("Error connecting to Couchbase:" , e);
+    } catch (IOException e) {
+      logger.error("Error connecting to Couchbase:", e);
       DTThrowable.rethrow(e);
     }
   }
@@ -241,15 +237,13 @@ public class CouchBaseStore implements Connectable
     CouchbaseClient clientPartition = null;
     try {
       nodes.add(new URI("http",urlString,"/pools", null, null));
-    }
-    catch (URISyntaxException ex) {
+    } catch (URISyntaxException ex) {
       DTThrowable.rethrow(ex);
     }
     try {
       clientPartition = new CouchbaseClient(nodes, bucket, password);
-    }
-    catch (IOException e) {
-     logger.error("Error connecting to Couchbase:" , e);
+    } catch (IOException e) {
+      logger.error("Error connecting to Couchbase:", e);
       DTThrowable.rethrow(e);
     }
     return clientPartition;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java
index 0c57ca6..2f0ce59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchBaseWindowStore.java
@@ -16,19 +16,23 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
-import java.io.*;
+import java.io.ByteArrayInputStream;
+import java.io.ByteArrayOutputStream;
+import java.io.DataInputStream;
+import java.io.DataOutputStream;
+import java.io.IOException;
 import java.util.concurrent.ExecutionException;
 import java.util.concurrent.TimeUnit;
 
-import com.couchbase.client.CouchbaseClient;
-import com.couchbase.client.CouchbaseConnectionFactoryBuilder;
-
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.datatorrent.lib.db.TransactionableStore;
+import org.apache.apex.malhar.lib.db.TransactionableStore;
+
+import com.couchbase.client.CouchbaseClient;
+import com.couchbase.client.CouchbaseConnectionFactoryBuilder;
 
 import com.datatorrent.netlet.util.DTThrowable;
 
@@ -92,9 +96,8 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
       cfb.setOpTimeout(timeout);  // wait up to 10 seconds for an operation to succeed
       cfb.setOpQueueMaxBlockTime(blockTime); // wait up to 10 second when trying to enqueue an operation
       clientMeta = new CouchbaseClient(cfb.buildCouchbaseConnection(baseURIs, bucketMeta, passwordMeta));
-    }
-    catch (IOException e) {
-      logger.error("Error connecting to Couchbase: " , e);
+    } catch (IOException e) {
+      logger.error("Error connecting to Couchbase: ", e);
       DTThrowable.rethrow(e);
     }
   }
@@ -119,11 +122,9 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
     String key = appId + "_" + operatorId + "_" + lastWindowValue;
     try {
       clientMeta.set(key, WindowIdBytes).get();
-    }
-    catch (InterruptedException ex) {
+    } catch (InterruptedException ex) {
       DTThrowable.rethrow(ex);
-    }
-    catch (ExecutionException ex) {
+    } catch (ExecutionException ex) {
       DTThrowable.rethrow(ex);
     }
 
@@ -163,8 +164,7 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
     try {
       result = dos.readLong();
       dos.close();
-    }
-    catch (IOException e) {
+    } catch (IOException e) {
       DTThrowable.rethrow(e);
     }
     return result;
@@ -179,8 +179,7 @@ public class CouchBaseWindowStore extends CouchBaseStore implements Transactiona
       dos.writeLong(l);
       result = baos.toByteArray();
       dos.close();
-    }
-    catch (IOException e) {
+    } catch (IOException e) {
       DTThrowable.rethrow(e);
     }
     return result;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java
index 5221b02..64ce9e2 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/CouchbasePOJOSetOperator.java
@@ -16,18 +16,18 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
 
 import java.util.ArrayList;
 
 import javax.validation.constraints.NotNull;
 
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
 
 /**
  * An implementation of Couchbase Output Operator which takes a POJO,serializes it into key,value
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java
index 1a6d05b..7b7dc59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchbase/package-info.java
@@ -17,4 +17,4 @@
  * under the License.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.couchbase;
+package org.apache.apex.malhar.contrib.couchbase;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java
index 1935e5c..ad3e797 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBInputOperator.java
@@ -16,20 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
 
 import java.io.IOException;
 import java.util.List;
 
 import javax.validation.constraints.Min;
 
-import com.google.common.base.Throwables;
-
 import org.ektorp.ViewQuery;
 import org.ektorp.ViewResult;
 
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
 
+import com.google.common.base.Throwables;
 import com.datatorrent.api.Context;
 
 
@@ -92,8 +91,7 @@ public abstract class AbstractCouchDBInputOperator<T> extends AbstractStoreInput
         T tuple = getTuple(row);
         outputPort.emit(tuple);
       }
-    }
-    catch (Throwable cause) {
+    } catch (Throwable cause) {
       Throwables.propagate(cause);
     }
     if (rows.size() > 0) {
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java
index 5e6c24b..8080eda 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractCouchDBOutputOperator.java
@@ -16,10 +16,9 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
 
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
-import java.util.Map;
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
 
 /**
  * Generic base output adaptor which saves tuples in the CouchDb.&nbsp; Subclasses should provide implementation for getting Document Id. <br/>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java
index 94c814f..e39a91c 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/AbstractMapBasedInputOperator.java
@@ -16,15 +16,15 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchdb;
-
-import org.codehaus.jackson.map.ObjectMapper;
-import org.ektorp.ViewResult;
+package org.apache.apex.malhar.contrib.couchdb;
 
 import java.io.IOException;
 import java.util.HashMap;
 import java.util.Map;
 
+import org.codehaus.jackson.map.ObjectMapper;
+import org.ektorp.ViewResult;
+
 /**
  * A base CouchDb input adaptor that emits a map. <br/>
  * <p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java
index 8111727..6d5defa 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
 
 import java.io.IOException;
 import java.util.ArrayList;
@@ -24,14 +24,18 @@ import java.util.List;
 
 import javax.validation.constraints.NotNull;
 
-import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.codehaus.jackson.JsonNode;
 import org.codehaus.jackson.map.ObjectMapper;
 import org.ektorp.ViewQuery;
 import org.ektorp.ViewResult.Row;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Setter;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterBoolean;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterDouble;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterInt;
+import org.apache.apex.malhar.lib.util.PojoUtils.SetterLong;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
 
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.*;
 import com.datatorrent.api.Context.OperatorContext;
 
 /**
@@ -184,8 +188,7 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
     try {
       // This code will be replaced after integration of creating POJOs on the fly utility.
       objectClass = Class.forName(outputClass);
-    }
-    catch (ClassNotFoundException ex) {
+    } catch (ClassNotFoundException ex) {
       throw new RuntimeException(ex);
     }
 
@@ -197,18 +200,15 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
       Class<?> type = null;
       try {
         type = objectClass.getDeclaredField(columns.get(i)).getType();
-      }
-      catch (NoSuchFieldException ex) {
+      } catch (NoSuchFieldException ex) {
         throw new RuntimeException(ex);
-      }
-      catch (SecurityException ex) {
+      } catch (SecurityException ex) {
         throw new RuntimeException(ex);
       }
       fieldType.add(type);
       if (type.isPrimitive()) {
         setterDoc.add(PojoUtils.constructSetter(objectClass, expressions.get(i), type));
-      }
-      else {
+      } else {
         setterDoc.add(PojoUtils.createSetter(objectClass, expressions.get(i), type));
       }
     }
@@ -222,11 +222,9 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
     Object obj;
     try {
       obj = objectClass.newInstance();
-    }
-    catch (InstantiationException ex) {
+    } catch (InstantiationException ex) {
       throw new RuntimeException(ex);
-    }
-    catch (IllegalAccessException ex) {
+    } catch (IllegalAccessException ex) {
       throw new RuntimeException(ex);
     }
 
@@ -240,21 +238,16 @@ public class CouchDBPOJOInputOperator extends AbstractCouchDBInputOperator<Objec
       if (type.isPrimitive()) {
         if (type == int.class) {
           ((SetterInt)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getIntValue());
-        }
-        else if (type == boolean.class) {
+        } else if (type == boolean.class) {
           ((SetterBoolean)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getBooleanValue());
-        }
-        else if (type == long.class) {
+        } else if (type == long.class) {
           ((SetterLong)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getLongValue());
-        }
-        else if (type == double.class) {
+        } else if (type == double.class) {
           ((SetterDouble)setterDoc.get(i)).set(obj, val.get(columns.get(i)).getDoubleValue());
-        }
-        else {
+        } else {
           throw new RuntimeException("Type is not supported");
         }
-      }
-      else {
+      } else {
         ((Setter<Object, Object>)setterDoc.get(i)).set(obj, mapper.readValue(val.get(columns.get(i)), type));
       }
     }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java
index f75473c..9d75272 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDBPOJOOutputOperator.java
@@ -16,15 +16,14 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
 
 import javax.validation.constraints.NotNull;
 
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
 import org.apache.hadoop.classification.InterfaceStability.Evolving;
 
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
-
 /**
  * Implementation of {@link AbstractCouchDBOutputOperator} that saves a POJO in the couch database. <br/>
  * <p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java
index 88c2eab..31e5423 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/CouchDbStore.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
 
 import java.io.IOException;
 import java.net.MalformedURLException;
@@ -32,7 +32,7 @@ import org.ektorp.http.HttpClient;
 import org.ektorp.http.StdHttpClient;
 import org.ektorp.impl.StdCouchDbInstance;
 
-import com.datatorrent.lib.db.Connectable;
+import org.apache.apex.malhar.lib.db.Connectable;
 
 /**
  * Implements a CouchDb store. <br/>
@@ -141,8 +141,7 @@ public class CouchDbStore implements Connectable
   {
     if (docId != null && dbConnector.contains(docId)) {
       dbConnector.update(document);
-    }
-    else {
+    } else {
       //create a document & if docId is null then couch db will generate a random id.
       dbConnector.create(document);
     }
@@ -166,8 +165,7 @@ public class CouchDbStore implements Connectable
     if (dbUrl != null) {
       try {
         builder.url(dbUrl);
-      }
-      catch (MalformedURLException e) {
+      } catch (MalformedURLException e) {
         throw new IllegalArgumentException(e.getMessage());
       }
     }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java
index f83f93c..6c8b813 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/MapBasedCouchDbOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
 
 import java.util.Map;
 
@@ -34,7 +34,7 @@ public class MapBasedCouchDbOutputOperator extends AbstractCouchDBOutputOperator
   @Override
   public String getDocumentId(Map<Object, Object> tuple)
   {
-    return (String) tuple.get("_id");
+    return (String)tuple.get("_id");
   }
 
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java
index be6ec34..13b7e17 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/couchdb/package-info.java
@@ -17,4 +17,4 @@
  * under the License.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.couchdb;
+package org.apache.apex.malhar.contrib.couchdb;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java
index 022530b..5d9e915 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchInputOperator.java
@@ -16,14 +16,15 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
 
 import org.elasticsearch.action.search.SearchRequestBuilder;
 import org.elasticsearch.action.search.SearchResponse;
 import org.elasticsearch.search.SearchHit;
 
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
+
 import com.datatorrent.api.Context.OperatorContext;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
 
 /**
  * This is the base implementation for a non transactional input operator for ElasticSearch
@@ -53,7 +54,7 @@ public abstract class AbstractElasticSearchInputOperator<T, S extends ElasticSea
   /**
    * Initializing transient fields such as ElasticSearchConnectable, SearchRequestBuilder
    *
-   * @see com.datatorrent.lib.db.AbstractStoreInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
+   * @see org.apache.apex.malhar.lib.db.AbstractStoreInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
    */
   @Override
   public void setup(OperatorContext t1)
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java
index 0282ae8..34b4440 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/AbstractElasticSearchOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
 
 import java.util.Queue;
 import java.util.concurrent.ArrayBlockingQueue;
@@ -27,9 +27,10 @@ import org.elasticsearch.action.bulk.BulkRequestBuilder;
 import org.elasticsearch.action.bulk.BulkResponse;
 import org.elasticsearch.action.index.IndexRequestBuilder;
 
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
+
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
 
 /**
  * This is the base implementation for a non-transactional batch output operator for ElasticSearch.
@@ -70,7 +71,7 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
   /**
    * Initialize transient fields such as {@code tupleBatch}
    *
-   * @see com.datatorrent.lib.db.AbstractStoreOutputOperator#setup(com.datatorrent.api.Context.OperatorContext)
+   * @see org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator#setup(com.datatorrent.api.Context.OperatorContext)
    */
   @Override
   public void setup(OperatorContext context)
@@ -83,7 +84,7 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
    * Adds tuple to the queue.
    * Calls {@link #processBatch()} if queue is full
    *
-   * @see com.datatorrent.lib.db.AbstractStoreOutputOperator#processTuple(java.lang.Object)
+   * @see org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator#processTuple(java.lang.Object)
    */
   public void processTuple(T tuple)
   {
@@ -128,7 +129,8 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
    * @param tuple
    * @return
    */
-  protected IndexRequestBuilder getIndexRequestBuilder(T tuple){
+  protected IndexRequestBuilder getIndexRequestBuilder(T tuple)
+  {
     IndexRequestBuilder indexRequestBuilder = new IndexRequestBuilder(store.client, getIndexName(tuple));
     String id = getId(tuple);
     if (id != null) {
@@ -176,22 +178,22 @@ public abstract class AbstractElasticSearchOutputOperator<T, S extends ElasticSe
    * @param tuple
    * @return
    */
-    protected abstract String getType(T tuple);
+  protected abstract String getType(T tuple);
 
     /**
      * @return the batchSize
      */
-    public int getBatchSize()
-    {
-      return batchSize;
-    }
+  public int getBatchSize()
+  {
+    return batchSize;
+  }
 
     /**
      * @param batchSize the batchSize to set
      */
-    public void setBatchSize(int batchSize)
-    {
-      this.batchSize = batchSize;
-    }
+  public void setBatchSize(int batchSize)
+  {
+    this.batchSize = batchSize;
+  }
 
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java
index 34eca95..c607393 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchConnectable.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
 
 import java.io.IOException;
 
@@ -25,7 +25,7 @@ import javax.validation.constraints.NotNull;
 import org.elasticsearch.client.transport.TransportClient;
 import org.elasticsearch.common.transport.InetSocketTransportAddress;
 
-import com.datatorrent.lib.db.Connectable;
+import org.apache.apex.malhar.lib.db.Connectable;
 
 /**
  * Elastic search base connector which has basic information for an operator <br>
@@ -91,7 +91,7 @@ public class ElasticSearchConnectable implements Connectable
   /*
    * (non-Javadoc)
    *
-   * @see com.datatorrent.lib.db.Connectable#connect()
+   * @see org.apache.apex.malhar.lib.db.Connectable#connect()
    */
   @Override
   public void connect() throws IOException
@@ -103,7 +103,7 @@ public class ElasticSearchConnectable implements Connectable
   /*
    * (non-Javadoc)
    *
-   * @see com.datatorrent.lib.db.Connectable#disconnect()
+   * @see org.apache.apex.malhar.lib.db.Connectable#disconnect()
    */
   @Override
   public void disconnect() throws IOException
@@ -116,7 +116,7 @@ public class ElasticSearchConnectable implements Connectable
   /*
    * (non-Javadoc)
    *
-   * @see com.datatorrent.lib.db.Connectable#isConnected()
+   * @see org.apache.apex.malhar.lib.db.Connectable#isConnected()
    */
   @Override
   public boolean isConnected()
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java
index dcbee9d..6774f10 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
 
 import java.util.Map;
 
@@ -50,7 +50,7 @@ public abstract class ElasticSearchMapInputOperator<T extends Map<String, Object
   /**
    * {@link SearchRequestBuilder} properties which do not change for each window are set during operator initialization.
    *
-   * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
+   * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchInputOperator#setup(com.datatorrent.api.Context.OperatorContext)
    */
   @Override
   public void setup(OperatorContext t1)
@@ -63,14 +63,14 @@ public abstract class ElasticSearchMapInputOperator<T extends Map<String, Object
    * (non-Javadoc)
    *
    * @see
-   * com.datatorrent.contrib.elasticsearch.AbstractElasticSearchInputOperator#convertToTuple(org.elasticsearch.search
+   * org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchInputOperator#convertToTuple(org.elasticsearch.search
    * .SearchHit)
    */
   @Override
   protected T convertToTuple(SearchHit hit)
   {
     Map<String, Object> tuple = hit.getSource();
-    return (T) tuple;
+    return (T)tuple;
   }
 
   /**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java
index 8616938..d85f0a3 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchMapOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
 
 import java.util.Map;
 
@@ -50,7 +50,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
    * (non-Javadoc)
    *
    * @see
-   * com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#setSource(org.elasticsearch.action.index
+   * org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#setSource(org.elasticsearch.action.index
    * .IndexRequestBuilder, java.lang.Object)
    */
   @Override
@@ -62,7 +62,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
   /*
    * (non-Javadoc)
    *
-   * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getId(java.lang.Object)
+   * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getId(java.lang.Object)
    */
   @Override
   protected String getId(T tuple)
@@ -104,7 +104,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
   /*
    * (non-Javadoc)
    *
-   * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getIndexName(java.lang.Object)
+   * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getIndexName(java.lang.Object)
    */
   @Override
   protected String getIndexName(T tuple)
@@ -122,7 +122,7 @@ public class ElasticSearchMapOutputOperator<T extends Map<String, Object>> exten
   }
 
   /* (non-Javadoc)
-   * @see com.datatorrent.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getType(java.lang.Object)
+   * @see org.apache.apex.malhar.contrib.elasticsearch.AbstractElasticSearchOutputOperator#getType(java.lang.Object)
    */
   @Override
   protected String getType(T tuple)
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java
index 9f7b66c..614fd59 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
 
 import java.io.IOException;
 
@@ -24,9 +24,9 @@ import javax.validation.constraints.NotNull;
 
 import org.elasticsearch.action.percolate.PercolateResponse;
 
-import com.datatorrent.common.util.BaseOperator;
 import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.api.DefaultOutputPort;
+import com.datatorrent.common.util.BaseOperator;
 import com.datatorrent.netlet.util.DTThrowable;
 
 /**
@@ -50,8 +50,8 @@ public class ElasticSearchPercolatorOperator extends BaseOperator
   protected transient ElasticSearchPercolatorStore store;
   public final transient DefaultOutputPort<PercolateResponse> outputPort = new DefaultOutputPort<PercolateResponse>();
 
-  public final transient DefaultInputPort<Object> inputPort = new DefaultInputPort<Object>() {
-
+  public final transient DefaultInputPort<Object> inputPort = new DefaultInputPort<Object>()
+  {
     /*
      * (non-Javadoc)
      *
@@ -61,7 +61,7 @@ public class ElasticSearchPercolatorOperator extends BaseOperator
     public void process(Object tuple)
     {
 
-      PercolateResponse response = store.percolate(new String[] { indexName }, documentType, tuple);
+      PercolateResponse response = store.percolate(new String[] {indexName}, documentType, tuple);
       outputPort.emit(response);
     }
   };
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java
index c13c025..19639c7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/ElasticSearchPercolatorStore.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
 
 import java.io.IOException;
 
@@ -58,7 +58,8 @@ public class ElasticSearchPercolatorStore extends ElasticSearchConnectable
     }
   }
 
-  public PercolateResponse percolate(String[] indexNames, String documentType, Object tuple){
+  public PercolateResponse percolate(String[] indexNames, String documentType, Object tuple)
+  {
     XContentBuilder docBuilder;
     try {
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java
index d89d53f..1c74c32 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/elasticsearch/package-info.java
@@ -17,4 +17,4 @@
  * under the License.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.elasticsearch;
+package org.apache.apex.malhar.contrib.elasticsearch;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java
index c377b96..b7e8820 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/AbstractEnricher.java
@@ -16,12 +16,17 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.io.IOException;
 import java.util.ArrayList;
 import java.util.List;
 
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
+import org.apache.apex.malhar.lib.db.cache.CacheStore;
+import org.apache.apex.malhar.lib.db.cache.CacheStore.ExpiryType;
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.FieldInfo.SupportType;
 import org.apache.hadoop.classification.InterfaceStability;
 
 import com.esotericsoftware.kryo.NotNull;
@@ -29,11 +34,6 @@ import com.esotericsoftware.kryo.NotNull;
 import com.datatorrent.api.Context;
 import com.datatorrent.api.Operator;
 import com.datatorrent.common.util.BaseOperator;
-import com.datatorrent.lib.db.cache.CacheManager;
-import com.datatorrent.lib.db.cache.CacheStore;
-import com.datatorrent.lib.db.cache.CacheStore.ExpiryType;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.FieldInfo.SupportType;
 
 /**
  * Base class for Enrichment Operator.&nbsp; Subclasses should provide implementation to getKey and convert.
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java
index 845a756..62d11a0 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/BackendLoader.java
@@ -16,15 +16,14 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.util.List;
 
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
+import org.apache.apex.malhar.lib.util.FieldInfo;
 import org.apache.hadoop.classification.InterfaceStability;
 
-import com.datatorrent.lib.db.cache.CacheManager;
-import com.datatorrent.lib.util.FieldInfo;
-
 /**
  * Interface for store to be used in enrichment
  *
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java
index 3121cf1..aa4a8bd 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/DelimitedFSLoader.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.io.IOException;
 import java.util.List;
@@ -30,14 +30,13 @@ import org.supercsv.cellprocessor.ift.CellProcessor;
 import org.supercsv.io.CsvMapReader;
 import org.supercsv.prefs.CsvPreference;
 
+import org.apache.apex.malhar.contrib.parser.CellProcessorBuilder;
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema;
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema.Field;
+import org.apache.apex.malhar.lib.util.ReusableStringReader;
 import org.apache.commons.lang.StringUtils;
 import org.apache.hadoop.classification.InterfaceStability;
 
-import com.datatorrent.contrib.parser.CellProcessorBuilder;
-import com.datatorrent.contrib.parser.DelimitedSchema;
-import com.datatorrent.contrib.parser.DelimitedSchema.Field;
-import com.datatorrent.lib.util.ReusableStringReader;
-
 /**
  * This implementation of {@link FSLoader} is used to load data from delimited
  * file.User needs to provide a schema as a string specified in a json format as
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java
index e04d6c4..cc23cb2 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FSLoader.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.io.BufferedReader;
 import java.io.IOException;
@@ -28,6 +28,8 @@ import java.util.Map;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
+import org.apache.apex.malhar.lib.util.FieldInfo;
 import org.apache.hadoop.classification.InterfaceStability;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FSDataInputStream;
@@ -38,8 +40,6 @@ import com.esotericsoftware.kryo.NotNull;
 import com.google.common.collect.Maps;
 
 import com.datatorrent.api.Component;
-import com.datatorrent.lib.db.cache.CacheManager;
-import com.datatorrent.lib.util.FieldInfo;
 
 /**
  * This implementation of {@link BackendLoader} loads the data from a given file
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java
index 2f54f66..e855c07 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/FixedWidthFSLoader.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.text.ParseException;
 import java.text.SimpleDateFormat;
@@ -30,6 +30,8 @@ import javax.validation.constraints.NotNull;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.contrib.parser.AbstractCsvParser.FIELD_TYPE;
+import org.apache.apex.malhar.contrib.parser.AbstractCsvParser.Field;
 import org.apache.commons.lang3.StringUtils;
 import org.apache.hadoop.classification.InterfaceStability;
 
@@ -39,9 +41,6 @@ import com.univocity.parsers.fixed.FixedWidthFields;
 import com.univocity.parsers.fixed.FixedWidthParser;
 import com.univocity.parsers.fixed.FixedWidthParserSettings;
 
-import com.datatorrent.contrib.parser.AbstractCsvParser.FIELD_TYPE;
-import com.datatorrent.contrib.parser.AbstractCsvParser.Field;
-
 /**
  * This implementation of {@link FSLoader} is used to load data from fixed width
  * file.User needs to set {@link FixedWidthFSLoader#fieldDescription} to specify
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java
index ce02c08..0ed9e22 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JDBCLoader.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.sql.PreparedStatement;
 import java.sql.ResultSet;
@@ -26,12 +26,12 @@ import java.util.ArrayList;
 import java.util.List;
 import java.util.Map;
 
+import org.apache.apex.malhar.lib.db.jdbc.JdbcStore;
+import org.apache.apex.malhar.lib.util.FieldInfo;
 import org.apache.commons.collections.CollectionUtils;
 import org.apache.hadoop.classification.InterfaceStability;
 
 import com.google.common.collect.Lists;
-import com.datatorrent.lib.db.jdbc.JdbcStore;
-import com.datatorrent.lib.util.FieldInfo;
 
 /**
  * <p>HBaseLoader extends from {@link JdbcStore} uses JDBC to connect and implements BackendLoaders interface.</p> <br/>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java
index 183d07a..1347100 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/JsonFSLoader.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.io.IOException;
 import java.util.Map;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java
index 6e7b5f7..d3d8df4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/MapEnricher.java
@@ -16,21 +16,21 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.util.ArrayList;
 import java.util.List;
 import java.util.Map;
 
+import org.apache.apex.malhar.lib.util.FieldInfo;
 import org.apache.hadoop.classification.InterfaceStability;
 
 import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.lib.util.FieldInfo;
 
 /**
  * This class takes a HashMap tuple as input and extracts value of the lookupKey configured
- * for this operator. It perform a lookup using {@link com.datatorrent.lib.db.cache.CacheManager} to
+ * for this operator. It perform a lookup using {@link org.apache.apex.malhar.lib.db.cache.CacheManager} to
  * find a matching entry and adds the result to the original tuple.
  *
  * <p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java
index 6d42fec..74f3f97 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/NullValuesCacheManager.java
@@ -16,12 +16,11 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
+import org.apache.apex.malhar.lib.db.cache.CacheManager;
 import org.apache.hadoop.classification.InterfaceStability;
 
-import com.datatorrent.lib.db.cache.CacheManager;
-
 /**
  * Null Values Cache Manager. Using this NULL entries can be specified explicitly.
  *
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java
index f22acdb..1f979ee 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/POJOEnricher.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.lang.reflect.Field;
 import java.util.ArrayList;
@@ -26,6 +26,9 @@ import java.util.Map;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
 import org.apache.commons.lang3.ClassUtils;
 import org.apache.hadoop.classification.InterfaceStability;
 
@@ -35,13 +38,11 @@ import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.api.DefaultOutputPort;
 import com.datatorrent.api.annotation.InputPortFieldAnnotation;
 import com.datatorrent.api.annotation.OutputPortFieldAnnotation;
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
 
 
 /**
  * This class takes a POJO as input and extracts the value of the lookupKey configured
- * for this operator. It perform a lookup using {@link com.datatorrent.lib.db.cache.CacheManager} to
+ * for this operator. It perform a lookup using {@link org.apache.apex.malhar.lib.db.cache.CacheManager} to
  * find a matching entry and adds the result to the original tuple.
  *
  * <p>
@@ -212,7 +213,7 @@ public class POJOEnricher extends AbstractEnricher<Object, Object>
 
   @SuppressWarnings({ "unchecked", "rawtypes" })
   private PojoUtils.Setter generateSettersForField(Class<?> klass, String outputFieldName)
-      throws NoSuchFieldException, SecurityException
+    throws NoSuchFieldException, SecurityException
   {
     Field f = klass.getDeclaredField(outputFieldName);
     Class c = ClassUtils.primitiveToWrapper(f.getType());
@@ -221,7 +222,7 @@ public class POJOEnricher extends AbstractEnricher<Object, Object>
 
   @SuppressWarnings({ "unchecked", "rawtypes" })
   private PojoUtils.Getter generateGettersForField(Class<?> klass, String inputFieldName)
-      throws NoSuchFieldException, SecurityException
+    throws NoSuchFieldException, SecurityException
   {
     Field f = klass.getDeclaredField(inputFieldName);
     Class c = ClassUtils.primitiveToWrapper(f.getType());
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java
index c6afbe1..0ae3ee7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/enrich/ReadOnlyBackup.java
@@ -16,15 +16,14 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.enrich;
+package org.apache.apex.malhar.contrib.enrich;
 
 import java.util.List;
 import java.util.Map;
 
+import org.apache.apex.malhar.lib.util.FieldInfo;
 import org.apache.hadoop.classification.InterfaceStability;
 
-import com.datatorrent.lib.util.FieldInfo;
-
 /**
  * ReadOnly abstract implementation of BackendLoader.
  *
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java
index 2bd0e67..3469039 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/formatter/CsvFormatter.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.formatter;
+package org.apache.apex.malhar.contrib.formatter;
 
 import java.io.IOException;
 import java.io.StringWriter;
@@ -34,14 +34,15 @@ import org.supercsv.io.CsvBeanWriter;
 import org.supercsv.io.ICsvBeanWriter;
 import org.supercsv.prefs.CsvPreference;
 
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema;
+import org.apache.apex.malhar.contrib.parser.DelimitedSchema.Field;
+import org.apache.apex.malhar.contrib.parser.Schema.FieldType;
+import org.apache.apex.malhar.lib.formatter.Formatter;
+
 import com.google.common.annotations.VisibleForTesting;
 
 import com.datatorrent.api.AutoMetric;
 import com.datatorrent.api.Context;
-import com.datatorrent.contrib.parser.DelimitedSchema;
-import com.datatorrent.contrib.parser.DelimitedSchema.Field;
-import com.datatorrent.contrib.parser.Schema.FieldType;
-import com.datatorrent.lib.formatter.Formatter;
 import com.datatorrent.netlet.util.DTThrowable;
 
 /**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java
index 497e6e4..d56d471 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeInputOperator.java
@@ -16,9 +16,9 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
 
-import com.datatorrent.lib.db.AbstractKeyValueStoreInputOperator;
+import org.apache.apex.malhar.lib.db.AbstractKeyValueStoreInputOperator;
 
 /**
  * This is the base implementation used for geode input adapters.&nbsp; A
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java
index dd0bad2..d3951fc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/AbstractGeodeOutputOperator.java
@@ -16,9 +16,9 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
 
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
 
 /**
  * This is the base implementation of geode output operators.&nbsp; A concrete
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java
index 2152b97..0c01373 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeCheckpointStore.java
@@ -16,9 +16,21 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
+
+import java.io.ByteArrayInputStream;
+import java.io.IOException;
+import java.io.Serializable;
+import java.util.ArrayList;
+import java.util.Iterator;
+import java.util.List;
+import java.util.Map;
+import java.util.Map.Entry;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.StorageAgentKeyValueStore;
 
-import com.datatorrent.lib.util.StorageAgentKeyValueStore;
 import com.esotericsoftware.kryo.Kryo;
 import com.esotericsoftware.kryo.io.Input;
 import com.esotericsoftware.kryo.io.Output;
@@ -32,17 +44,6 @@ import com.gemstone.gemfire.cache.query.Query;
 import com.gemstone.gemfire.cache.query.QueryService;
 import com.gemstone.gemfire.cache.query.SelectResults;
 import com.google.common.collect.Maps;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import java.io.ByteArrayInputStream;
-import java.io.IOException;
-import java.io.Serializable;
-import java.util.ArrayList;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-import java.util.Map.Entry;
 
 /**
  * Geode Store implementation of {@link StorageAgentKeyValueStore} Uses {@link Kryo}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java
index 691c2c1..819d531 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeKeyValueStorageAgent.java
@@ -16,14 +16,13 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
 
 import java.io.Serializable;
 
+import org.apache.apex.malhar.lib.util.AbstractKeyValueStorageAgent;
 import org.apache.hadoop.conf.Configuration;
 
-import com.datatorrent.lib.util.AbstractKeyValueStorageAgent;
-
 /**
  * Storage Agent implementation which uses {@link GeodeCheckpointStore} for operator
  * checkpointing
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java
index c7d22c7..d2001e5 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodePOJOOutputOperator.java
@@ -16,15 +16,14 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
 
+import org.apache.apex.malhar.lib.util.FieldInfo;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.TableInfo;
 import org.apache.hadoop.classification.InterfaceStability.Evolving;
 
-import com.datatorrent.lib.util.FieldInfo;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
-import com.datatorrent.lib.util.TableInfo;
-
 /**
  *
  * @displayName Geode Output Operator
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java
index d345661..803a83e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/GeodeStore.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
 
 import java.io.IOException;
 import java.io.Serializable;
@@ -27,7 +27,7 @@ import java.util.Map;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import org.apache.accumulo.core.client.impl.thrift.ThriftTest.Processor.throwsError;
+import org.apache.apex.malhar.lib.db.KeyValueStore;
 
 import com.gemstone.gemfire.cache.CacheClosedException;
 import com.gemstone.gemfire.cache.CacheWriterException;
@@ -43,8 +43,6 @@ import com.gemstone.gemfire.cache.query.QueryInvocationTargetException;
 import com.gemstone.gemfire.cache.query.SelectResults;
 import com.gemstone.gemfire.cache.query.TypeMismatchException;
 
-import com.datatorrent.lib.db.KeyValueStore;
-
 /**
  * Provides the implementation of a Geode store.
  * Geode is a distributed in-memory database
@@ -145,7 +143,7 @@ public class GeodeStore implements KeyValueStore, Serializable
     if (region == null) {
       region = clientCache.getRegion(regionName);
       if (region == null) {
-        region = clientCache.<Object, Object> createClientRegionFactory(ClientRegionShortcut.PROXY).create(regionName);
+        region = clientCache.<Object, Object>createClientRegionFactory(ClientRegionShortcut.PROXY).create(regionName);
       }
     }
 
@@ -164,7 +162,7 @@ public class GeodeStore implements KeyValueStore, Serializable
     region = clientCache.getRegion(getRegionName());
 
     if (region == null) {
-      region = clientCache.<Object, Object> createClientRegionFactory(ClientRegionShortcut.PROXY).create(
+      region = clientCache.<Object, Object>createClientRegionFactory(ClientRegionShortcut.PROXY).create(
           getRegionName());
     }
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java
index 9e948c4..d7aa9d5 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/geode/RegionCreateFunction.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.geode;
+package org.apache.apex.malhar.contrib.geode;
 
 import java.util.ArrayList;
 import java.util.List;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java
index a2c1297..d83768d 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseAppendOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 
@@ -29,7 +29,7 @@ import org.apache.hadoop.hbase.client.HTable;
 import com.datatorrent.netlet.util.DTThrowable;
 
 /**
- * A base implementation of a StoreOutputOperator operator that stores tuples in HBase columns and offers 
+ * A base implementation of a StoreOutputOperator operator that stores tuples in HBase columns and offers
  * non-transactional append.&nbsp; Subclasses should provide implementation for appending operations. <br>
  * <p>
  * <br>
@@ -49,16 +49,19 @@ import com.datatorrent.netlet.util.DTThrowable;
  *            The tuple type
  * @since 1.0.2
  */
-public abstract class AbstractHBaseAppendOutputOperator<T> extends AbstractHBaseOutputOperator<T> {
+public abstract class AbstractHBaseAppendOutputOperator<T> extends AbstractHBaseOutputOperator<T>
+{
   private static final transient Logger logger = LoggerFactory
       .getLogger(AbstractHBaseAppendOutputOperator.class);
 
-  public AbstractHBaseAppendOutputOperator() {
+  public AbstractHBaseAppendOutputOperator()
+  {
     store = new HBaseStore();
   }
 
   @Override
-  public void processTuple(T tuple, HTable table) {
+  public void processTuple(T tuple, HTable table)
+  {
     Append append = operationAppend(tuple);
     try {
       table.append(append);
diff --git a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseOutputOperator.java
similarity index 93%
rename from contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseOutputOperator.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseOutputOperator.java
index b9d3c49..f07f123 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseOutputOperator.java
@@ -16,15 +16,15 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
+import org.apache.apex.malhar.lib.db.AbstractStoreOutputOperator;
 import org.apache.hadoop.hbase.client.HTable;
 
 import com.datatorrent.api.Operator;
-import com.datatorrent.lib.db.AbstractStoreOutputOperator;
 
 /**
- * A base implementation of a HBase output operator that stores tuples in HBase and offers 
+ * A base implementation of a HBase output operator that stores tuples in HBase and offers
  * non-transactional Insert.&nbsp; Subclasses should provide implementation for specific Hbase operations.
  *
  * @since 3.8.0
@@ -32,7 +32,7 @@ import com.datatorrent.lib.db.AbstractStoreOutputOperator;
 public abstract class AbstractHBaseOutputOperator<T> extends AbstractStoreOutputOperator<T, HBaseStore> implements OutputAdapter.Driver<T>, Operator.CheckpointNotificationListener
 {
   private transient OutputAdapter<T> outputAdapter;
-  
+
   public AbstractHBaseOutputOperator()
   {
     outputAdapter = new OutputAdapter<T>(store, this);
@@ -43,13 +43,13 @@ public abstract class AbstractHBaseOutputOperator<T> extends AbstractStoreOutput
   {
     outputAdapter.processTuple(tuple);
   }
-  
+
   @Override
   public abstract void processTuple(T tuple, HTable table);
 
   /**
    * Get the table name for tuple.
-   * 
+   *
    * Implementations can override this method to return the table name where the tuple should be written to.
    * Return null to write to default table
    * @param tuple The tuple
@@ -84,5 +84,5 @@ public abstract class AbstractHBaseOutputOperator<T> extends AbstractStoreOutput
   {
 
   }
-  
+
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java
index 2421936..cc5f8e4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBasePutOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.InterruptedIOException;
 
@@ -48,7 +48,8 @@ import com.datatorrent.netlet.util.DTThrowable;
  *            The tuple type
  * @since 1.0.2
  */
-public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOutputOperator<T> {
+public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOutputOperator<T>
+{
   private static final transient Logger logger = LoggerFactory.getLogger(AbstractHBasePutOutputOperator.class);
 
   public AbstractHBasePutOutputOperator()
@@ -56,7 +57,6 @@ public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOut
     store = new HBaseStore();
   }
 
-  @Override
   public void processTuple(T tuple, HTable table)
   {
     Put put = operationPut(tuple);
@@ -72,5 +72,5 @@ public abstract class AbstractHBasePutOutputOperator<T> extends AbstractHBaseOut
   }
 
   public abstract Put operationPut(T t);
-  
+
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java
index 2fa2146..1747cde 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowAppendOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 
@@ -30,7 +30,9 @@ import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.netlet.util.DTThrowable;
 
 /**
- * A base implementation of an AggregateTransactionableStoreOutputOperator operator that stores tuples in HBase columns and provides batch append.&nbsp; Subclasses should provide implementation for appending operations. <br>
+ * A base implementation of an AggregateTransactionableStoreOutputOperator
+ * operator that stores tuples in HBase columns and provides batch append.&nbsp;
+ * Subclasses should provide implementation for appending operations. <br>
  * <p>
  * <br>
  * This class provides a HBase output operator that can be used to store tuples
@@ -45,17 +47,17 @@ import com.datatorrent.netlet.util.DTThrowable;
  * guarantee each tuple is written only once to HBase in case the operator is
  * restarted from an earlier checkpoint. It only tries to minimize the number of
  * duplicates limiting it to the tuples that were processed in the window when
- * the operator shutdown.
- * It supports atleast once and atmost once processing modes.
- * Exactly once is not supported
+ * the operator shutdown. It supports atleast once and atmost once processing
+ * modes. Exactly once is not supported
+ *
  * @displayName Abstract HBase Window Append Output
  * @category Output
  * @tags hbase, append, transactionable, batch
- * @param <T>
- *            The tuple type
+ * @param <T> The tuple type
  * @since 1.0.2
  */
-public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T> {
+public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T>
+{
   private static final transient Logger logger = LoggerFactory.getLogger(AbstractHBaseWindowAppendOutputOperator.class);
   private transient ProcessingMode mode;
 
@@ -74,12 +76,14 @@ public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends Abstrac
     this.mode = mode;
   }
 
-  public AbstractHBaseWindowAppendOutputOperator() {
+  public AbstractHBaseWindowAppendOutputOperator()
+  {
     store = new HBaseWindowStore();
   }
 
   @Override
-  public void processTuple(T tuple, HTable table) {
+  public void processTuple(T tuple, HTable table)
+  {
     try {
       Append append = operationAppend(tuple);
       table.append(append);
@@ -103,8 +107,8 @@ public abstract class AbstractHBaseWindowAppendOutputOperator<T> extends Abstrac
   @Override
   public void setup(OperatorContext context)
   {
-    mode=context.getValue(OperatorContext.PROCESSING_MODE);
-    if(mode==ProcessingMode.EXACTLY_ONCE){
+    mode = context.getValue(OperatorContext.PROCESSING_MODE);
+    if (mode == ProcessingMode.EXACTLY_ONCE) {
       throw new RuntimeException("This operator only supports atmost once and atleast once processing modes");
     }
     super.setup(context);
diff --git a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseWindowOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowOutputOperator.java
similarity index 95%
rename from contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseWindowOutputOperator.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowOutputOperator.java
index 665b463..fb21bf9 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/hbase/AbstractHBaseWindowOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowOutputOperator.java
@@ -16,15 +16,15 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.db.AbstractPassThruTransactionableStoreOutputOperator;
 import org.apache.hadoop.hbase.client.HTable;
 
 import com.datatorrent.api.Operator;
-import com.datatorrent.lib.db.AbstractPassThruTransactionableStoreOutputOperator;
 
 /**
  * A base implementation of an AggregateTransactionableStoreOutputOperator that stores tuples in HBase
@@ -38,7 +38,7 @@ public abstract class AbstractHBaseWindowOutputOperator<T> extends AbstractPassT
    * Both at checkpoint window and end window, flush the tuples as application window may not align with
    * end window especially when it is more than one streaming window
    */
-  
+
   private static final Logger logger = LoggerFactory.getLogger(AbstractHBaseWindowOutputOperator.class);
 
   private transient OutputAdapter<T> outputAdapter;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java
index 70fbeeb..e6c65fc 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/AbstractHBaseWindowPutOutputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 
@@ -30,7 +30,9 @@ import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.netlet.util.DTThrowable;
 
 /**
- * A base implementation of an AggregateTransactionableStoreOutputOperator operator for storing tuples in HBase rows and provides a batch put.Subclasses should provide implementation for put operation. <br>
+ * A base implementation of an AggregateTransactionableStoreOutputOperator
+ * operator for storing tuples in HBase rows and provides a batch put.Subclasses
+ * should provide implementation for put operation. <br>
  * <p>
  * <br>
  * This class provides a HBase output operator that can be used to store tuples
@@ -45,20 +47,20 @@ import com.datatorrent.netlet.util.DTThrowable;
  * guarantee each tuple is written only once to HBase in case the operator is
  * restarted from an earlier checkpoint. It only tries to minimize the number of
  * duplicates limiting it to the tuples that were processed in the window when
- * the operator shutdown.
- * It supports atleast once and atmost once processing modes.
- * Exactly once is not supported
+ * the operator shutdown. It supports atleast once and atmost once processing
+ * modes. Exactly once is not supported
+ *
  * @displayName Abstract HBase Window Put Output
  * @category Output
  * @tags hbase, put, transactionable, batch
- * @param <T>
- *            The tuple type
+ * @param <T> The tuple type
  * @since 1.0.2
  */
-public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T> {
+public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHBaseWindowOutputOperator<T>
+{
   private static final transient Logger logger = LoggerFactory.getLogger(AbstractHBaseWindowPutOutputOperator.class);
   private transient ProcessingMode mode;
-  
+
   @Deprecated
   public ProcessingMode getMode()
   {
@@ -71,12 +73,14 @@ public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHB
     this.mode = mode;
   }
 
-  public AbstractHBaseWindowPutOutputOperator() {
+  public AbstractHBaseWindowPutOutputOperator()
+  {
     store = new HBaseWindowStore();
   }
 
   @Override
-  public void processTuple(T tuple, HTable table) {
+  public void processTuple(T tuple, HTable table)
+  {
     try {
       Put put = operationPut(tuple);
       table.put(put);
@@ -86,14 +90,13 @@ public abstract class AbstractHBaseWindowPutOutputOperator<T> extends AbstractHB
     }
   }
 
-
   public abstract Put operationPut(T t) throws IOException;
 
   @Override
   public void setup(OperatorContext context)
   {
-    mode=context.getValue(context.PROCESSING_MODE);
-    if(mode==ProcessingMode.EXACTLY_ONCE){
+    mode = context.getValue(context.PROCESSING_MODE);
+    if (mode == ProcessingMode.EXACTLY_ONCE) {
       throw new RuntimeException("This operator only supports atmost once and atleast once processing modes");
     }
     super.setup(context);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java
index 7420569..468dd8c 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseCsvMappingPutOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 import java.io.StringReader;
@@ -25,27 +25,31 @@ import java.util.HashMap;
 import java.util.Map;
 import java.util.Map.Entry;
 
-import org.apache.hadoop.hbase.client.Put;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import org.supercsv.io.CsvListReader;
 import org.supercsv.io.ICsvListReader;
 import org.supercsv.prefs.CsvPreference;
+import org.apache.apex.malhar.lib.util.ReusableStringReader;
+import org.apache.hadoop.hbase.client.Put;
 
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.util.ReusableStringReader;
 
 /**
- * Takes a configuration string which tells us about the position of the row, or column.&nbsp; The incoming tuples are inserted accordingly.
+ * Takes a configuration string which tells us about the position of the row, or
+ * column.&nbsp; The incoming tuples are inserted accordingly.
  * <p>
+ *
  * @displayName HBase Csv Mapping Put
  * @category Output
  * @tags hbase, csv, put, String
  * @since 1.0.4
  */
-public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOperator<String> {
-  private class ColDef {
+public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOperator<String>
+{
+  private class ColDef
+  {
     String colFam;
     String colName;
   }
@@ -58,23 +62,26 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
   private transient ArrayList<String> csvLineList = new ArrayList<String>();
   private String mappingString;
 
-  public void setMappingString(String mappingString) {
+  public void setMappingString(String mappingString)
+  {
     this.mappingString = mappingString;
   }
 
   @Override
-  public Put operationPut(String t) throws IOException {
+  public Put operationPut(String t) throws IOException
+  {
     return parseLine(t);
   }
 
-  public void parseMapping() {
+  public void parseMapping()
+  {
     ICsvListReader listReader = null;
     StringReader sr = null;
     ArrayList<String> csvList = new ArrayList<String>();
     try {
       sr = new StringReader(mappingString);
       listReader = new CsvListReader(sr,CsvPreference.STANDARD_PREFERENCE);
-      csvList = (ArrayList<String>) listReader.read();
+      csvList = (ArrayList<String>)listReader.read();
     } catch (IOException e) {
       logger.error("Cannot read the mapping string", e);
       DTThrowable.rethrow(e);
@@ -89,9 +96,9 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
     }
     for (int index = 0; index < csvList.size(); index++) {
       String value = csvList.get(index);
-      if (value.equals("row"))
+      if (value.equals("row")) {
         rowIndex = index;
-      else {
+      } else {
         ColDef c = new ColDef();
         c.colFam = value.substring(0, value.indexOf('.'));
         c.colName = value.substring(value.indexOf('.') + 1);
@@ -100,11 +107,12 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
     }
   }
 
-  public Put parseLine(String s) {
+  public Put parseLine(String s)
+  {
     Put put = null;
     try {
       lineSr.open(s);
-      csvLineList = (ArrayList<String>) lineListReader.read();
+      csvLineList = (ArrayList<String>)lineListReader.read();
     } catch (IOException e) {
       logger.error("Cannot read the property string", e);
       DTThrowable.rethrow(e);
@@ -120,15 +128,17 @@ public class HBaseCsvMappingPutOperator extends AbstractHBaseWindowPutOutputOper
   }
 
   @Override
-  public void setup(OperatorContext context) {
+  public void setup(OperatorContext context)
+  {
     super.setup(context);
     parseMapping();
     lineListReader = new CsvListReader(lineSr,
-        CsvPreference.STANDARD_PREFERENCE);
+    CsvPreference.STANDARD_PREFERENCE);
   }
 
   @Override
-  public void teardown() {
+  public void teardown()
+  {
     super.teardown();
     try {
       lineSr.close();
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java
index 6a34a91..23fa2c5 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldInfo.java
@@ -16,96 +16,83 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
+import org.apache.apex.malhar.lib.util.FieldInfo;
 import org.apache.hadoop.hbase.util.Bytes;
 
-import com.datatorrent.lib.util.FieldInfo;
-
 /**
  * @since 3.1.0
  */
 
 public class HBaseFieldInfo extends FieldInfo
 {
-	private String familyName;
-
-	public HBaseFieldInfo()
-	{
-	}
-
-	public HBaseFieldInfo( String columnName, String columnExpression, SupportType type, String familyName )
-	{
-	  super( columnName, columnExpression, type );
-	  setFamilyName( familyName );
-	}
-
-	public String getFamilyName()
-	{
-		return familyName;
-	}
-
-	public void setFamilyName(String familyName)
-	{
-		this.familyName = familyName;
-	}
+  private String familyName;
 
-	public byte[] toBytes( Object value )
-	{
-		final SupportType type = getType();
-		switch( type )
-		{
-		case BOOLEAN:
-		  return Bytes.toBytes( (Boolean)value );
-
-		case SHORT:
-		  return Bytes.toBytes( (Short)value );
-
-		case INTEGER:
-		  return Bytes.toBytes( (Integer)value );
-
-		case LONG:
-		  return Bytes.toBytes( (Long)value );
+  public HBaseFieldInfo()
+  {
+  }
 
-		case FLOAT:
-		  return Bytes.toBytes( (Float)value );
+  public HBaseFieldInfo( String columnName, String columnExpression, SupportType type, String familyName )
+  {
+    super( columnName, columnExpression, type );
+    setFamilyName( familyName );
+  }
 
-		case DOUBLE:
-		  return Bytes.toBytes( (Double)value );
+  public String getFamilyName()
+  {
+    return familyName;
+  }
 
-		case STRING:
-		  return Bytes.toBytes( (String)value );
-		}
-		throw new IllegalArgumentException( "Unsupported type: " + type );
-	}
+  public void setFamilyName(String familyName)
+  {
+    this.familyName = familyName;
+  }
 
-	public Object toValue( byte[] bytes )
-	{
+  public byte[] toBytes( Object value )
+  {
     final SupportType type = getType();
-    switch( type )
-    {
-    case BOOLEAN:
-      return Bytes.toBoolean( bytes );
-
-    case SHORT:
-      return Bytes.toShort( bytes );
-
-    case INTEGER:
-      return Bytes.toInt( bytes );
-
-    case LONG:
-      return Bytes.toLong( bytes );
-
-    case FLOAT:
-      return Bytes.toFloat( bytes );
-
-    case DOUBLE:
-      return Bytes.toDouble( bytes );
+    switch (type) {
+      case BOOLEAN:
+        return Bytes.toBytes( (Boolean)value );
+      case SHORT:
+        return Bytes.toBytes( (Short)value );
+      case INTEGER:
+        return Bytes.toBytes( (Integer)value );
+      case LONG:
+        return Bytes.toBytes( (Long)value );
+      case FLOAT:
+        return Bytes.toBytes( (Float)value );
+      case DOUBLE:
+        return Bytes.toBytes( (Double)value );
+      case STRING:
+        return Bytes.toBytes( (String)value );
+      default:
+        throw new IllegalArgumentException("Unsupported type: " + type);
+    }
+  }
 
-    case STRING:
-      return Bytes.toString( bytes );
+  public Object toValue( byte[] bytes )
+  {
+    final SupportType type = getType();
+    switch (type) {
+      case BOOLEAN:
+        return Bytes.toBoolean( bytes );
+      case SHORT:
+        return Bytes.toShort( bytes );
+      case INTEGER:
+        return Bytes.toInt( bytes );
+      case LONG:
+        return Bytes.toLong( bytes );
+      case FLOAT:
+        return Bytes.toFloat( bytes );
+      case DOUBLE:
+        return Bytes.toDouble( bytes );
+      case STRING:
+        return Bytes.toString( bytes );
+      default:
+        throw new IllegalArgumentException("Unsupported type: " + type);
     }
-    throw new IllegalArgumentException( "Unsupported type: " + type );
   }
 
 }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java
index fea56a0..a49399e 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseFieldValueGenerator.java
@@ -16,12 +16,12 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.util.List;
 
-import com.datatorrent.lib.util.FieldValueGenerator;
-import com.datatorrent.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.FieldValueGenerator;
+import org.apache.apex.malhar.lib.util.PojoUtils;
 
 /**
  * A {@link FieldValueGenerator} implementation for {@link HBaseFieldInfo}
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java
index 37270d5..12608c8 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseGetOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.client.Get;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java
index 6f11621..82e50b4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseInputOperator.java
@@ -16,10 +16,11 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
+
+import org.apache.apex.malhar.lib.db.AbstractStoreInputOperator;
 
 import com.datatorrent.api.InputOperator;
-import com.datatorrent.lib.db.AbstractStoreInputOperator;
 
 /**
  * A base implementation of hbase input operator which derives from HBaseOperatorBase. <br>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java
index d35281c..5fd97e1 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseNameValueCsvPutOperator.java
@@ -16,45 +16,48 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
+
+import java.io.IOException;
+import java.io.StringReader;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.Map;
+import javax.validation.constraints.NotNull;
 
-import com.datatorrent.api.Context.OperatorContext;
-import com.datatorrent.netlet.util.DTThrowable;
-import com.datatorrent.lib.util.ReusableStringReader;
-import org.apache.hadoop.hbase.client.Put;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import org.supercsv.io.CsvListReader;
 import org.supercsv.io.ICsvListReader;
 import org.supercsv.prefs.CsvPreference;
+import org.apache.apex.malhar.lib.util.ReusableStringReader;
+import org.apache.hadoop.hbase.client.Put;
 
-import javax.validation.constraints.NotNull;
-import java.io.IOException;
-import java.io.StringReader;
-import java.util.ArrayList;
-import java.util.HashMap;
-import java.util.Map;
+import com.datatorrent.api.Context.OperatorContext;
+import com.datatorrent.netlet.util.DTThrowable;
 
 /**
- * An implementation of HBase window put operator that inserts a string of key value pairs which are mapped to corresponding rows, columns.
+ * An implementation of HBase window put operator that inserts a string of key
+ * value pairs which are mapped to corresponding rows, columns.
  * <p>
  * Accepts a string of key value pairs containing the data to be inserted.These
  * are mapped to corresponding rows,column families and columns using a property
  * file and inserted into hbase Example: input string will be of the form
- * name="xyz", st="patrickhenry" ,ct="fremont", sa="california", the properties will contain
- * properties of form name=row, sa=address.street, ct=address.city, sa=address.state.
- * With the above mapping a row xyz is created. The value patrickhenry is inserted into
- * columnfamily address and column street of row xyz. Other values are inserted
- * similarly.
+ * name="xyz", st="patrickhenry" ,ct="fremont", sa="california", the properties
+ * will contain properties of form name=row, sa=address.street, ct=address.city,
+ * sa=address.state. With the above mapping a row xyz is created. The value
+ * patrickhenry is inserted into columnfamily address and column street of row
+ * xyz. Other values are inserted similarly.
  *
  * @displayName HBase Name Value Csv Put
  * @category Output
  * @tags csv, hbase, put
  * @since 1.0.2
  */
-public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOperator<String> {
-
-  private class ColDef {
+public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOperator<String>
+{
+  private class ColDef
+  {
     String colFam;
     String colName;
   }
@@ -66,26 +69,29 @@ public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOp
   private transient Map<String, ColDef> colMap = new HashMap<String, ColDef>();
   private transient Map<String, String> linemap = new HashMap<String, String>();
   private transient ICsvListReader lineListReader = null;
-  private transient ReusableStringReader lineSr=new ReusableStringReader();
+  private transient ReusableStringReader lineSr = new ReusableStringReader();
   private transient ArrayList<String> csvLineList = new ArrayList<String>();
 
-  public void setMapping(String mapping) {
+  public void setMapping(String mapping)
+  {
     this.mapping = mapping;
   }
 
   @Override
-  public Put operationPut(String t) {
+  public Put operationPut(String t)
+  {
     return parseLine(t);
   }
 
-  public void parseMapping() {
+  public void parseMapping()
+  {
     ICsvListReader listReader = null;
-    StringReader sr=null;
+    StringReader sr = null;
     ArrayList<String> csvList = new ArrayList<String>();
     try {
-      sr=new StringReader(mapping);
+      sr = new StringReader(mapping);
       listReader = new CsvListReader(sr,CsvPreference.STANDARD_PREFERENCE);
-      csvList = (ArrayList<String>) listReader.read();
+      csvList = (ArrayList<String>)listReader.read();
     } catch (IOException e) {
       logger.error("Cannot read the mapping string", e);
       DTThrowable.rethrow(e);
@@ -110,11 +116,12 @@ public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOp
     }
   }
 
-  public Put parseLine(String s) {
+  public Put parseLine(String s)
+  {
     Put put = null;
     try {
       lineSr.open(s);
-      csvLineList = (ArrayList<String>) lineListReader.read();
+      csvLineList = (ArrayList<String>)lineListReader.read();
     } catch (IOException e) {
       logger.error("Cannot read the property string", e);
       DTThrowable.rethrow(e);
@@ -144,14 +151,16 @@ public class HBaseNameValueCsvPutOperator extends AbstractHBaseWindowPutOutputOp
   }
 
   @Override
-  public void setup(OperatorContext context) {
+  public void setup(OperatorContext context)
+  {
     super.setup(context);
     parseMapping();
     lineListReader = new CsvListReader(lineSr,CsvPreference.STANDARD_PREFERENCE);
   }
 
   @Override
-  public void teardown() {
+  public void teardown()
+  {
     super.teardown();
     try {
       lineSr.close();
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java
index 8bed824..04a51b4 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseOperatorBase.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 import org.apache.hadoop.conf.Configuration;
@@ -103,7 +103,8 @@ public class HBaseOperatorBase
    * Get the configuration.
    * @return The configuration
    */
-  public Configuration getConfiguration() {
+  public Configuration getConfiguration()
+  {
     return configuration;
   }
 
@@ -113,7 +114,8 @@ public class HBaseOperatorBase
    * should be called before using the configuration or table.
    * @throws IOException
    */
-  protected void setupConfiguration() throws IOException {
+  protected void setupConfiguration() throws IOException
+  {
     configuration = HBaseConfiguration.create();
     configuration.set("hbase.zookeeper.quorum", zookeeperQuorum);
     configuration.set("hbase.zookeeper.property.clientPort", "" + zookeeperClientPort);
@@ -127,7 +129,8 @@ public class HBaseOperatorBase
    * @return The HBase table if configuration setup was successful, null otherwise
    * @throws IOException
    */
-  protected HTable getTable() throws IOException {
+  protected HTable getTable() throws IOException
+  {
     return table;
   }
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java
index e459ec7..cb36dbf 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOInputOperator.java
@@ -16,12 +16,18 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.util.List;
 
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
+
+import org.apache.apex.malhar.lib.util.FieldValueGenerator;
+import org.apache.apex.malhar.lib.util.FieldValueGenerator.ValueConverter;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Setter;
+import org.apache.apex.malhar.lib.util.TableInfo;
 import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.CellUtil;
@@ -29,14 +35,8 @@ import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.util.Bytes;
 
-import com.datatorrent.lib.util.FieldValueGenerator;
-import com.datatorrent.lib.util.FieldValueGenerator.ValueConverter;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Setter;
-import com.datatorrent.lib.util.TableInfo;
 import com.datatorrent.api.Context;
 import com.datatorrent.api.DefaultOutputPort;
-import com.datatorrent.api.Context.OperatorContext;
 
 /**
  * HBasePOJOInputOperator reads data from a HBase store, converts it to a POJO and puts it on the output port.
@@ -96,20 +96,20 @@ public class HBasePOJOInputOperator extends HBaseScanOperator<Object>
   {
     try {
       String readRow = Bytes.toString(result.getRow());
-      if( readRow.equals( getLastReadRow() )) {
+      if ( readRow.equals( getLastReadRow() )) {
         return null;
       }
 
       Object instance = pojoType.newInstance();
       rowSetter.set(instance, readRow);
 
-       List<Cell> cells = result.listCells();
-       for (Cell cell : cells) {
-         String columnName = Bytes.toString(CellUtil.cloneQualifier(cell));
-         String columnFamily = Bytes.toString(CellUtil.cloneFamily(cell));
+      List<Cell> cells = result.listCells();
+      for (Cell cell : cells) {
+        String columnName = Bytes.toString(CellUtil.cloneQualifier(cell));
+        String columnFamily = Bytes.toString(CellUtil.cloneFamily(cell));
         byte[] value = CellUtil.cloneValue(cell);
-         ((HBaseFieldValueGenerator)fieldValueGenerator).setColumnValue(instance, columnName, columnFamily, value,
-             valueConverter);
+        ((HBaseFieldValueGenerator)fieldValueGenerator).setColumnValue(instance, columnName, columnFamily, value,
+            valueConverter);
       }
 
       setLastReadRow(readRow);
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java
index 37c962b..d70f3c7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBasePOJOPutOperator.java
@@ -16,20 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.util.List;
 
+import org.apache.apex.malhar.lib.util.FieldValueGenerator;
+import org.apache.apex.malhar.lib.util.FieldValueGenerator.FieldValueHandler;
+import org.apache.apex.malhar.lib.util.PojoUtils;
+import org.apache.apex.malhar.lib.util.PojoUtils.Getter;
+import org.apache.apex.malhar.lib.util.TableInfo;
 import org.apache.hadoop.classification.InterfaceStability.Evolving;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.util.Bytes;
 
-import com.datatorrent.lib.util.FieldValueGenerator;
-import com.datatorrent.lib.util.FieldValueGenerator.FieldValueHandler;
-import com.datatorrent.lib.util.PojoUtils;
-import com.datatorrent.lib.util.PojoUtils.Getter;
-import com.datatorrent.lib.util.TableInfo;
-
 /**
  * @displayName HBase Output Operator
  * @category Output
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java
index b694e67..f525058 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseScanOperator.java
@@ -16,36 +16,38 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 import java.util.Queue;
-import org.apache.hadoop.hbase.KeyValue;
-import org.apache.hadoop.hbase.client.Scan;
+
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
+import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.ResultScanner;
+import org.apache.hadoop.hbase.client.Scan;
 
+import com.google.common.collect.Queues;
 import com.datatorrent.api.AutoMetric;
 import com.datatorrent.api.Context;
-import com.datatorrent.api.Operator;
 import com.datatorrent.api.Context.OperatorContext;
-import com.google.common.collect.Queues;
-
-import org.apache.hadoop.hbase.client.HTable;
-import org.apache.hadoop.hbase.client.Result;
-import org.apache.hadoop.hbase.client.ResultScanner;
+import com.datatorrent.api.Operator;
 
 /**
- * A base implementation of hbase input operator that retrieves tuples from HBase columns and provides scan operation. <br>
+ * A base implementation of hbase input operator that retrieves tuples from
+ * HBase columns and provides scan operation. <br>
  * <p>
  * <br>
- * This class provides a HBase input operator that can be used to retrieve tuples from rows in a
- * HBase table. The class should be extended by the end-operator developer. The extending class should
- * implement operationScan and getTuple methods. The operationScan method should provide a HBase Scan
- * metric object that specifies where to retrieve the tuple information from the table. The getTuple method
- * should map the contents of a Result from the Scan result to a tuple.<br>
+ * This class provides a HBase input operator that can be used to retrieve
+ * tuples from rows in a HBase table. The class should be extended by the
+ * end-operator developer. The extending class should implement operationScan
+ * and getTuple methods. The operationScan method should provide a HBase Scan
+ * metric object that specifies where to retrieve the tuple information from the
+ * table. The getTuple method should map the contents of a Result from the Scan
+ * result to a tuple.<br>
  *
  * <br>
+ *
  * @displayName HBase Scan
  * @category Output
  * @tags hbase, scan, input operator
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java
index 2b0ffa7..d46bc6f 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseStore.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 import java.io.InterruptedIOException;
@@ -27,6 +27,7 @@ import javax.validation.constraints.Min;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.db.Connectable;
 import org.apache.commons.lang.ArrayUtils;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.HBaseConfiguration;
@@ -41,7 +42,6 @@ import com.google.common.cache.LoadingCache;
 import com.google.common.cache.RemovalListener;
 import com.google.common.cache.RemovalNotification;
 
-import com.datatorrent.lib.db.Connectable;
 /**
  * A {@link Connectable} that uses HBase to connect to stores and implements Connectable interface.
  * <p>
@@ -50,15 +50,15 @@ import com.datatorrent.lib.db.Connectable;
  * @tags store, hbase
  * @since 1.0.2
  */
-public class HBaseStore implements Connectable {
-
+public class HBaseStore implements Connectable
+{
   public static final String USER_NAME_SPECIFIER = "%USER_NAME%";
 
   private static final Logger logger = LoggerFactory.getLogger(HBaseStore.class);
 
   private String zookeeperQuorum;
   private int zookeeperClientPort;
-  
+
   // Default table name if specified
   protected String tableName;
 
@@ -67,7 +67,7 @@ public class HBaseStore implements Connectable {
   // Default interval 30 min
   protected long reloginCheckInterval = 30 * 60 * 1000;
   protected transient Thread loginRenewer;
-  private volatile transient boolean doRelogin;
+  private transient volatile boolean doRelogin;
 
   protected transient HTable table;
   // Multi - table
@@ -76,13 +76,14 @@ public class HBaseStore implements Connectable {
 
   @Min(1)
   protected int maxOpenTables = Integer.MAX_VALUE;
-  
+
   /**
    * Get the zookeeper quorum location.
    *
    * @return The zookeeper quorum location
    */
-  public String getZookeeperQuorum() {
+  public String getZookeeperQuorum()
+  {
     return zookeeperQuorum;
   }
 
@@ -92,7 +93,8 @@ public class HBaseStore implements Connectable {
    * @param zookeeperQuorum
    *            The zookeeper quorum location
    */
-  public void setZookeeperQuorum(String zookeeperQuorum) {
+  public void setZookeeperQuorum(String zookeeperQuorum)
+  {
     this.zookeeperQuorum = zookeeperQuorum;
   }
 
@@ -101,7 +103,8 @@ public class HBaseStore implements Connectable {
    *
    * @return The zookeeper client port
    */
-  public int getZookeeperClientPort() {
+  public int getZookeeperClientPort()
+  {
     return zookeeperClientPort;
   }
 
@@ -111,7 +114,8 @@ public class HBaseStore implements Connectable {
    * @param zookeeperClientPort
    *            The zookeeper client port
    */
-  public void setZookeeperClientPort(int zookeeperClientPort) {
+  public void setZookeeperClientPort(int zookeeperClientPort)
+  {
     this.zookeeperClientPort = zookeeperClientPort;
   }
 
@@ -120,7 +124,8 @@ public class HBaseStore implements Connectable {
    *
    * @return The HBase table name
    */
-  public String getTableName() {
+  public String getTableName()
+  {
     return tableName;
   }
 
@@ -130,7 +135,8 @@ public class HBaseStore implements Connectable {
    * @param tableName
    *            The HBase table name
    */
-  public void setTableName(String tableName) {
+  public void setTableName(String tableName)
+  {
     this.tableName = tableName;
   }
 
@@ -199,8 +205,8 @@ public class HBaseStore implements Connectable {
 
   /**
    * Gets the allowedtableNames
-   * 
-   * @return  allowedTableNames 
+   *
+   * @return  allowedTableNames
    */
   public String[] getAllowedTableNames()
   {
@@ -211,14 +217,15 @@ public class HBaseStore implements Connectable {
   {
     this.allowedTableNames = allowedTableNames;
   }
-  
+
   /**
    * Get the HBase table .
    *
    * @return The HBase table
    * @omitFromUI
    */
-  public HTable getTable() {
+  public HTable getTable()
+  {
     return table;
   }
 
@@ -230,16 +237,18 @@ public class HBaseStore implements Connectable {
    * @return The HBase table
    * @omitFromUI
    */
-  public HTable getTable(String tableName) {
-    if ((tableName == null) || tableName.equals(this.tableName))
+  public HTable getTable(String tableName)
+  {
+    if ((tableName == null) || tableName.equals(this.tableName)) {
       return getTable();
+    }
     try {
       return tableCache.get(tableName);
     } catch (Exception e) {
       throw Throwables.propagate(e);
     }
   }
-  
+
   public void flushTables() throws InterruptedIOException, RetriesExhaustedWithDetailsException
   {
     if (table != null) {
@@ -249,7 +258,7 @@ public class HBaseStore implements Connectable {
       flushTable(entry.getValue());
     }
   }
-  
+
   protected void flushTable(HTable table) throws InterruptedIOException, RetriesExhaustedWithDetailsException
   {
     table.flushCommits();
@@ -260,7 +269,8 @@ public class HBaseStore implements Connectable {
    *
    * @return The configuration
    */
-  public Configuration getConfiguration() {
+  public Configuration getConfiguration()
+  {
     return configuration;
   }
 
@@ -270,14 +280,16 @@ public class HBaseStore implements Connectable {
    * @param configuration
    *            The configuration
    */
-  public void setConfiguration(Configuration configuration) {
+  public void setConfiguration(Configuration configuration)
+  {
     this.configuration = configuration;
   }
 
   protected transient Configuration configuration;
 
   @Override
-  public void connect() throws IOException {
+  public void connect() throws IOException
+  {
     if ((principal != null) && (keytabPath != null)) {
       String lprincipal = evaluateProperty(principal);
       String lkeytabPath = evaluateProperty(keytabPath);
@@ -317,12 +329,12 @@ public class HBaseStore implements Connectable {
     if (zookeeperClientPort != 0) {
       configuration.set("hbase.zookeeper.property.clientPort", "" + zookeeperClientPort);
     }
-    
+
     // Connect to default table if specified
     if (tableName != null) {
       table = connectTable(tableName);
     }
-    
+
     CacheLoader<String, HTable> cacheLoader = new CacheLoader<String, HTable>()
     {
       @Override
@@ -331,7 +343,7 @@ public class HBaseStore implements Connectable {
         return loadTable(key);
       }
     };
-    
+
     RemovalListener<String, HTable> removalListener = new RemovalListener<String, HTable>()
     {
       @Override
@@ -340,9 +352,9 @@ public class HBaseStore implements Connectable {
         unloadTable(notification.getValue());
       }
     };
-    
+
     int maxCacheSize = (tableName == null) ? maxOpenTables : (maxOpenTables - 1);
-    
+
     tableCache = CacheBuilder.<String, HTable>newBuilder().maximumSize(maxCacheSize).removalListener(removalListener).build(cacheLoader);
   }
 
@@ -362,7 +374,7 @@ public class HBaseStore implements Connectable {
       logger.warn("Could not close table", e);
     }
   }
-  
+
   protected HTable connectTable(String tableName) throws IOException
   {
     HTable table = new HTable(configuration, tableName);
@@ -373,13 +385,14 @@ public class HBaseStore implements Connectable {
   private String evaluateProperty(String property) throws IOException
   {
     if (property.contains(USER_NAME_SPECIFIER)) {
-     property = property.replaceAll(USER_NAME_SPECIFIER, UserGroupInformation.getLoginUser().getShortUserName());
+      property = property.replaceAll(USER_NAME_SPECIFIER, UserGroupInformation.getLoginUser().getShortUserName());
     }
     return property;
   }
 
   @Override
-  public void disconnect() throws IOException {
+  public void disconnect() throws IOException
+  {
     if (loginRenewer != null) {
       doRelogin = false;
       loginRenewer.interrupt();
@@ -392,7 +405,8 @@ public class HBaseStore implements Connectable {
   }
 
   @Override
-  public boolean isConnected() {
+  public boolean isConnected()
+  {
     // not applicable to hbase
     return false;
   }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java
index 2c9f539..38d6195 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/HBaseWindowStore.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.IOException;
 import java.io.InterruptedIOException;
@@ -24,6 +24,7 @@ import java.io.InterruptedIOException;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import org.apache.apex.malhar.lib.db.TransactionableStore;
 import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HTableDescriptor;
 import org.apache.hadoop.hbase.KeyValue;
@@ -36,8 +37,8 @@ import org.apache.hadoop.hbase.util.Bytes;
 
 import com.google.common.base.Throwables;
 
-import com.datatorrent.lib.db.TransactionableStore;
 import com.datatorrent.netlet.util.DTThrowable;
+
 /**
  * <p>Provides transaction support to the operators by implementing TransactionableStore abstract methods. </p>
  * <p>
@@ -49,7 +50,8 @@ import com.datatorrent.netlet.util.DTThrowable;
  * @tags store, transactional
  * @since 1.0.2
  */
-public class HBaseWindowStore extends HBaseStore implements TransactionableStore {
+public class HBaseWindowStore extends HBaseStore implements TransactionableStore
+{
   private static final transient Logger logger = LoggerFactory.getLogger(HBaseWindowStore.class);
   private static final String DEFAULT_ROW_NAME = "HBaseOperator_row";
   private static final String DEFAULT_COLUMN_FAMILY_NAME = "HBaseOutputOperator_cf";
@@ -63,14 +65,16 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
   private transient String lastWindowColumnName;
   private transient byte[] lastWindowColumnBytes;
 
-  public HBaseWindowStore() {
+  public HBaseWindowStore()
+  {
     rowName = DEFAULT_ROW_NAME;
     columnFamilyName = DEFAULT_COLUMN_FAMILY_NAME;
     lastWindowColumnName = DEFAULT_LAST_WINDOW_PREFIX_COLUMN_NAME;
     constructKeys();
   }
 
-  private void constructKeys() {
+  private void constructKeys()
+  {
     rowBytes = Bytes.toBytes(rowName);
     columnFamilyBytes = Bytes.toBytes(columnFamilyName);
   }
@@ -80,7 +84,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
    *
    * @return The row name
    */
-  public String getRowName() {
+  public String getRowName()
+  {
     return rowName;
   }
 
@@ -90,7 +95,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
    * @param rowName
    *            The row name
    */
-  public void setRowName(String rowName) {
+  public void setRowName(String rowName)
+  {
     this.rowName = rowName;
     constructKeys();
   }
@@ -100,7 +106,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
    *
    * @return The column family name
    */
-  public String getColumnFamilyName() {
+  public String getColumnFamilyName()
+  {
     return columnFamilyName;
   }
 
@@ -110,13 +117,15 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
    * @param columnFamilyName
    *            The column family name
    */
-  public void setColumnFamilyName(String columnFamilyName) {
+  public void setColumnFamilyName(String columnFamilyName)
+  {
     this.columnFamilyName = columnFamilyName;
     constructKeys();
   }
 
   @Override
-  public void connect() throws IOException {
+  public void connect() throws IOException
+  {
     super.connect();
     HTableDescriptor tdesc = table.getTableDescriptor();
     if (!tdesc.hasFamily(columnFamilyBytes)) {
@@ -133,13 +142,14 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
   }
 
   @Override
-  public void beginTransaction() {
+  public void beginTransaction()
+  {
     // HBase does not support transactions so this method left empty
-
   }
 
   @Override
-  public void commitTransaction() {
+  public void commitTransaction()
+  {
     try {
       flushTables();
     } catch (InterruptedIOException | RetriesExhaustedWithDetailsException e) {
@@ -148,19 +158,21 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
   }
 
   @Override
-  public void rollbackTransaction() {
+  public void rollbackTransaction()
+  {
     // HBase does not support transactions so this method left empty
-
   }
 
   @Override
-  public boolean isInTransaction() {
+  public boolean isInTransaction()
+  {
     // HBase does not support transactions so this method left empty
     return false;
   }
 
   @Override
-  public long getCommittedWindowId(String appId, int operatorId) {
+  public long getCommittedWindowId(String appId, int operatorId)
+  {
     byte[] value = null;
     try {
       String columnKey = appId + "_" + operatorId + "_" + lastWindowColumnName;
@@ -183,12 +195,14 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
     if (value != null) {
       long longval = Bytes.toLong(value);
       return longval;
-    } else
+    } else {
       return -1;
+    }
   }
 
   @Override
-  public void storeCommittedWindowId(String appId, int operatorId,long windowId) {
+  public void storeCommittedWindowId(String appId, int operatorId,long windowId)
+  {
     byte[] WindowIdBytes = Bytes.toBytes(windowId);
     String columnKey = appId + "_" + operatorId + "_" + lastWindowColumnName;
     lastWindowColumnBytes = Bytes.toBytes(columnKey);
@@ -206,7 +220,8 @@ public class HBaseWindowStore extends HBaseStore implements TransactionableStore
   }
 
   @Override
-  public void removeCommittedWindowId(String appId, int operatorId) {
+  public void removeCommittedWindowId(String appId, int operatorId)
+  {
     // Not applicable to hbase
   }
 
diff --git a/contrib/src/main/java/com/datatorrent/contrib/hbase/OutputAdapter.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/OutputAdapter.java
similarity index 98%
rename from contrib/src/main/java/com/datatorrent/contrib/hbase/OutputAdapter.java
rename to contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/OutputAdapter.java
index 31317bd..e503ebb 100644
--- a/contrib/src/main/java/com/datatorrent/contrib/hbase/OutputAdapter.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/OutputAdapter.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
 
 import java.io.InterruptedIOException;
 
@@ -75,7 +75,9 @@ public class OutputAdapter<T>
   interface Driver<T>
   {
     void processTuple(T tuple, HTable table);
+
     String getTableName(T tuple);
+
     void errorTuple(T tuple);
   }
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java
index 0ccd6c2..2df5c10 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/hbase/package-info.java
@@ -20,4 +20,4 @@
  * HBase operators and utilities.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.hbase;
+package org.apache.apex.malhar.contrib.hbase;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java
index a6fb42f..997c890 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/PythonOperator.java
@@ -16,17 +16,23 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.jython;
-
-import com.datatorrent.api.Context.OperatorContext;
-import com.datatorrent.lib.script.ScriptOperator;
-import org.python.core.*;
-import org.python.util.PythonInterpreter;
+package org.apache.apex.malhar.contrib.jython;
 
 import java.util.HashMap;
 import java.util.Iterator;
 import java.util.Map;
 
+import org.python.core.PyCode;
+import org.python.core.PyIterator;
+import org.python.core.PyObject;
+import org.python.core.PyStringMap;
+import org.python.core.PyTuple;
+import org.python.util.PythonInterpreter;
+
+import org.apache.apex.malhar.lib.script.ScriptOperator;
+
+import com.datatorrent.api.Context.OperatorContext;
+
 /**
  * An operator that executes a python script and passes the input as bindings.
  * <p></p>
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java
index 08b3a3e..9fc7de7 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/jython/package-info.java
@@ -17,4 +17,4 @@
  * under the License.
  */
 @org.apache.hadoop.classification.InterfaceStability.Evolving
-package com.datatorrent.contrib.jython;
+package org.apache.apex.malhar.contrib.jython;
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java
index b11c8ef..ec082d1 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractExactlyOnceKafkaOutputOperator.java
@@ -16,12 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
 
+import java.nio.ByteBuffer;
+import java.util.HashMap;
+import java.util.Map;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import com.google.common.collect.Sets;
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultInputPort;
 import com.datatorrent.common.util.Pair;
-import com.google.common.collect.Sets;
 import kafka.api.FetchRequest;
 import kafka.api.FetchRequestBuilder;
 import kafka.javaapi.FetchResponse;
@@ -32,13 +39,6 @@ import kafka.message.Message;
 import kafka.message.MessageAndOffset;
 import kafka.producer.KeyedMessage;
 import kafka.producer.Partitioner;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import java.nio.ByteBuffer;
-import java.util.HashMap;
-import java.util.Map;
-
 
 /**
  * This is a base implementation of a Kafka output operator,
@@ -82,7 +82,6 @@ import java.util.Map;
  */
 public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends AbstractKafkaOutputOperator<K, V>
 {
-
   private Map<Integer, Pair<byte[], byte[]>>  lastMsgs;
 
   private transient  Partitioner partitioner;
@@ -94,9 +93,9 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
   {
     super.setup(context);
     try {
-      String className = (String) getConfigProperties().get(KafkaMetadataUtil.PRODUCER_PROP_PARTITIONER);
+      String className = (String)getConfigProperties().get(KafkaMetadataUtil.PRODUCER_PROP_PARTITIONER);
       if (className != null) {
-        partitioner = (Partitioner) Class.forName(className).newInstance();
+        partitioner = (Partitioner)Class.forName(className).newInstance();
       }
     } catch (Exception e) {
       throw new RuntimeException("Failed to initialize partitioner", e);
@@ -108,22 +107,22 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
   /**
    * This input port receives tuples that will be written out to Kafka.
    */
-  public final transient DefaultInputPort<T> inputPort = new DefaultInputPort<T>() {
+  public final transient DefaultInputPort<T> inputPort = new DefaultInputPort<T>()
+  {
     @Override
     public void process(T tuple)
     {
       Pair<K, V> keyValue = tupleToKeyValue(tuple);
       int pid = 0;
 
-      if(partitioner!=null){
+      if (partitioner != null) {
         pid = partitioner.partition(keyValue.first, partitionNum);
       }
 
       Pair<byte[], byte[]> lastMsg = lastMsgs.get(pid);
-
-      if(lastMsg == null || compareToLastMsg(keyValue, lastMsg) > 0){
+      if (lastMsg == null || compareToLastMsg(keyValue, lastMsg) > 0) {
         getProducer().send(new KeyedMessage<K, V>(getTopic(), keyValue.first, keyValue.second));
-        sendCount ++;
+        sendCount++;
       } else {
         // ignore tuple because kafka has already had the tuple
         logger.debug("Ingore tuple " + tuple);
@@ -134,7 +133,6 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
 
   private void initializeLastProcessingOffset()
   {
-
     // read last received kafka message
     TopicMetadata tm = KafkaMetadataUtil.getTopicMetadata(Sets.newHashSet((String)getConfigProperties().get(KafkaMetadataUtil.PRODUCER_PROP_BROKERLIST)), this.getTopic());
 
@@ -170,9 +168,7 @@ public abstract class AbstractExactlyOnceKafkaOutputOperator<T, K, V> extends Ab
         key.get(keyBytes);
         lastMsgs.put(pm.partitionId(), new Pair<byte[], byte[]>(keyBytes, valueBytes));
       }
-
     }
-
   }
 
   /**
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java
index 1218f4a..9130923 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaInputOperator.java
@@ -16,7 +16,35 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
+
+import java.io.IOException;
+import java.lang.reflect.Array;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+
+import javax.validation.Valid;
+import javax.validation.constraints.Min;
+import javax.validation.constraints.NotNull;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.apex.malhar.lib.util.KryoCloneUtils;
+import org.apache.apex.malhar.lib.wal.WindowDataManager;
+import org.apache.commons.lang3.StringUtils;
+import org.apache.commons.lang3.tuple.MutablePair;
+import org.apache.commons.lang3.tuple.Pair;
+
+import com.google.common.base.Joiner;
+import com.google.common.collect.Lists;
+import com.google.common.collect.Sets;
 
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.DefaultPartition;
@@ -28,10 +56,6 @@ import com.datatorrent.api.Stats;
 import com.datatorrent.api.StatsListener;
 import com.datatorrent.api.annotation.OperatorAnnotation;
 import com.datatorrent.api.annotation.Stateless;
-import com.datatorrent.lib.util.KryoCloneUtils;
-import com.google.common.base.Joiner;
-import com.google.common.collect.Lists;
-import com.google.common.collect.Sets;
 
 import kafka.api.FetchRequest;
 import kafka.api.FetchRequestBuilder;
@@ -42,30 +66,7 @@ import kafka.javaapi.consumer.SimpleConsumer;
 import kafka.message.Message;
 import kafka.message.MessageAndOffset;
 
-import org.apache.apex.malhar.lib.wal.WindowDataManager;
-import org.apache.commons.lang3.StringUtils;
-import org.apache.commons.lang3.tuple.MutablePair;
-import org.apache.commons.lang3.tuple.Pair;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import javax.validation.Valid;
-import javax.validation.constraints.Min;
-import javax.validation.constraints.NotNull;
-
-import java.io.IOException;
-import java.lang.reflect.Array;
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.HashMap;
-import java.util.HashSet;
-import java.util.Iterator;
-import java.util.LinkedList;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-
-import static com.datatorrent.contrib.kafka.KafkaConsumer.KafkaMeterStatsUtil.getOffsetsForPartitions;
+import static org.apache.apex.malhar.contrib.kafka.KafkaConsumer.KafkaMeterStatsUtil.getOffsetsForPartitions;
 
 /**
  * This is a base implementation of a Kafka input operator, which consumes data from Kafka message bus.&nbsp;
@@ -226,7 +227,8 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
    * processing of messages doesn't get stuck.
    * @return The maximum for the total size
      */
-  public long getMaxTotalMsgSizePerWindow() {
+  public long getMaxTotalMsgSizePerWindow()
+  {
     return maxTotalMsgSizePerWindow;
   }
 
@@ -236,7 +238,8 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
    *
    * @param maxTotalMsgSizePerWindow The maximum for the total size
      */
-  public void setMaxTotalMsgSizePerWindow(long maxTotalMsgSizePerWindow) {
+  public void setMaxTotalMsgSizePerWindow(long maxTotalMsgSizePerWindow)
+  {
     this.maxTotalMsgSizePerWindow = maxTotalMsgSizePerWindow;
   }
 
@@ -256,7 +259,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
     }
     this.context = context;
     operatorId = context.getId();
-    if(consumer instanceof HighlevelKafkaConsumer && !(windowDataManager instanceof WindowDataManager.NoopWindowDataManager)) {
+    if (consumer instanceof HighlevelKafkaConsumer && !(windowDataManager instanceof WindowDataManager.NoopWindowDataManager)) {
       throw new RuntimeException("Idempotency is not supported for High Level Kafka Consumer");
     }
     windowDataManager.setup(context);
@@ -289,7 +292,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
       if (recoveredData != null) {
         Map<String, List<PartitionMetadata>> pms = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().topic);
         if (pms != null) {
-          SimpleKafkaConsumer cons = (SimpleKafkaConsumer) getConsumer();
+          SimpleKafkaConsumer cons = (SimpleKafkaConsumer)getConsumer();
           // add all partition request in one Fretch request together
           FetchRequestBuilder frb = new FetchRequestBuilder().clientId(cons.getClientId());
           for (Map.Entry<KafkaPartition, MutablePair<Long, Integer>> rc : recoveredData.entrySet()) {
@@ -299,13 +302,14 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
             Iterator<PartitionMetadata> pmIterator = pmsVal.iterator();
             PartitionMetadata pm = pmIterator.next();
             while (pm.partitionId() != kp.getPartitionId()) {
-              if (!pmIterator.hasNext())
+              if (!pmIterator.hasNext()) {
                 break;
+              }
               pm = pmIterator.next();
             }
-            if (pm.partitionId() != kp.getPartitionId())
+            if (pm.partitionId() != kp.getPartitionId()) {
               continue;
-
+            }
             Broker bk = pm.leader();
 
             frb.addFetch(consumer.topic, rc.getKey().getPartitionId(), rc.getValue().left, cons.getBufferSize());
@@ -319,13 +323,14 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
               emitTuple(kafkaMessage);
               offsetStats.put(kp, msg.offset());
               count = count + 1;
-              if (count.equals(rc.getValue().right))
+              if (count.equals(rc.getValue().right)) {
                 break;
+              }
             }
           }
         }
       }
-      if(windowId == windowDataManager.getLargestCompletedWindow()) {
+      if (windowId == windowDataManager.getLargestCompletedWindow()) {
         // Start the consumer at the largest recovery window
         SimpleKafkaConsumer cons = (SimpleKafkaConsumer)getConsumer();
         // Set the offset positions to the consumer
@@ -337,8 +342,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
         cons.resetOffset(currentOffsets);
         cons.start();
       }
-    }
-    catch (IOException e) {
+    } catch (IOException e) {
       throw new RuntimeException("replay", e);
     }
   }
@@ -354,8 +358,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
     if (currentWindowId > windowDataManager.getLargestCompletedWindow()) {
       try {
         windowDataManager.save(currentWindowRecoveryState, currentWindowId);
-      }
-      catch (IOException e) {
+      } catch (IOException e) {
         throw new RuntimeException("saving recovery", e);
       }
     }
@@ -397,8 +400,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
 
     try {
       windowDataManager.committed(windowId);
-    }
-    catch (IOException e) {
+    } catch (IOException e) {
       throw new RuntimeException("deleting state", e);
     }
   }
@@ -453,10 +455,10 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
       emitTotalMsgSize += message.msg.size();
       offsetStats.put(message.kafkaPart, message.offSet);
       MutablePair<Long, Integer> offsetAndCount = currentWindowRecoveryState.get(message.kafkaPart);
-      if(offsetAndCount == null) {
+      if (offsetAndCount == null) {
         currentWindowRecoveryState.put(message.kafkaPart, new MutablePair<Long, Integer>(message.offSet, 1));
       } else {
-        offsetAndCount.setRight(offsetAndCount.right+1);
+        offsetAndCount.setRight(offsetAndCount.right + 1);
       }
     }
   }
@@ -507,7 +509,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
 
     boolean isInitialParitition = true;
     // check if it's the initial partition
-    if(partitions.iterator().hasNext()) {
+    if (partitions.iterator().hasNext()) {
       isInitialParitition = partitions.iterator().next().getStats() == null;
     }
 
@@ -516,7 +518,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
 
     // initialize the offset
     Map<KafkaPartition, Long> initOffset = null;
-    if(isInitialParitition && offsetManager !=null){
+    if (isInitialParitition && offsetManager != null) {
       initOffset = offsetManager.loadInitialOffsets();
       logger.info("Initial offsets: {} ", "{ " + Joiner.on(", ").useForNull("").withKeyValueSeparator(": ").join(initOffset) + " }");
     }
@@ -527,95 +529,92 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
 
     switch (strategy) {
 
-    // For the 1 to 1 mapping The framework will create number of operator partitions based on kafka topic partitions
-    // Each operator partition will consume from only one kafka partition
-    case ONE_TO_ONE:
-
-      if (isInitialParitition) {
-        lastRepartitionTime = System.currentTimeMillis();
-        logger.info("[ONE_TO_ONE]: Initializing partition(s)");
-        // get partition metadata for topics.
-        // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
-        // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
-        Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
-
-        // initialize the number of operator partitions according to number of kafka partitions
-
-        newPartitions = new LinkedList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>();
-        for (Map.Entry<String, List<PartitionMetadata>> kp : kafkaPartitions.entrySet()) {
-          String clusterId = kp.getKey();
-          for (PartitionMetadata pm : kp.getValue()) {
-            logger.info("[ONE_TO_ONE]: Create operator partition for cluster {}, topic {}, kafka partition {} ", clusterId, getConsumer().topic, pm.partitionId());
-            newPartitions.add(createPartition(Sets.newHashSet(new KafkaPartition(clusterId, consumer.topic, pm.partitionId())), initOffset));
+      // For the 1 to 1 mapping The framework will create number of operator partitions based on kafka topic partitions
+      // Each operator partition will consume from only one kafka partition
+      case ONE_TO_ONE:
+
+        if (isInitialParitition) {
+          lastRepartitionTime = System.currentTimeMillis();
+          logger.info("[ONE_TO_ONE]: Initializing partition(s)");
+          // get partition metadata for topics.
+          // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
+          // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
+          Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
+
+          // initialize the number of operator partitions according to number of kafka partitions
+
+          newPartitions = new LinkedList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>();
+          for (Map.Entry<String, List<PartitionMetadata>> kp : kafkaPartitions.entrySet()) {
+            String clusterId = kp.getKey();
+            for (PartitionMetadata pm : kp.getValue()) {
+              logger.info("[ONE_TO_ONE]: Create operator partition for cluster {}, topic {}, kafka partition {} ", clusterId, getConsumer().topic, pm.partitionId());
+              newPartitions.add(createPartition(Sets.newHashSet(new KafkaPartition(clusterId, consumer.topic, pm.partitionId())), initOffset));
+            }
+          }
+          resultPartitions = newPartitions;
+          numPartitionsChanged = true;
+        } else if (newWaitingPartition.size() != 0) {
+          // add partition for new kafka partition
+          for (KafkaPartition newPartition : newWaitingPartition) {
+            logger.info("[ONE_TO_ONE]: Add operator partition for cluster {}, topic {}, partition {}", newPartition.getClusterId(), getConsumer().topic, newPartition.getPartitionId());
+            partitions.add(createPartition(Sets.newHashSet(newPartition), null));
           }
+          newWaitingPartition.clear();
+          resultPartitions = partitions;
+          numPartitionsChanged = true;
         }
-        resultPartitions = newPartitions;
-        numPartitionsChanged = true;
-      }
-      else if (newWaitingPartition.size() != 0) {
-        // add partition for new kafka partition
-        for (KafkaPartition newPartition : newWaitingPartition) {
-          logger.info("[ONE_TO_ONE]: Add operator partition for cluster {}, topic {}, partition {}", newPartition.getClusterId(), getConsumer().topic, newPartition.getPartitionId());
-          partitions.add(createPartition(Sets.newHashSet(newPartition), null));
+        break;
+      // For the 1 to N mapping The initial partition number is defined by stream application
+      // Afterwards, the framework will dynamically adjust the partition and allocate consumers to as less operator partitions as it can
+      //  and guarantee the total intake rate for each operator partition is below some threshold
+      case ONE_TO_MANY:
+
+        if (getConsumer() instanceof HighlevelKafkaConsumer) {
+          throw new UnsupportedOperationException("[ONE_TO_MANY]: The high-level consumer is not supported for ONE_TO_MANY partition strategy.");
         }
-        newWaitingPartition.clear();
-        resultPartitions = partitions;
-        numPartitionsChanged = true;
-      }
-      break;
-    // For the 1 to N mapping The initial partition number is defined by stream application
-    // Afterwards, the framework will dynamically adjust the partition and allocate consumers to as less operator partitions as it can
-    //  and guarantee the total intake rate for each operator partition is below some threshold
-    case ONE_TO_MANY:
-
-      if (getConsumer() instanceof HighlevelKafkaConsumer) {
-        throw new UnsupportedOperationException("[ONE_TO_MANY]: The high-level consumer is not supported for ONE_TO_MANY partition strategy.");
-      }
 
-      if (isInitialParitition || newWaitingPartition.size() != 0) {
-        lastRepartitionTime = System.currentTimeMillis();
-        logger.info("[ONE_TO_MANY]: Initializing partition(s)");
-        // get partition metadata for topics.
-        // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
-        // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
-        Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
-
-        int size = initialPartitionCount;
-        @SuppressWarnings("unchecked")
-        Set<KafkaPartition>[] kps = (Set<KafkaPartition>[]) Array.newInstance((new HashSet<KafkaPartition>()).getClass(), size);
-        int i = 0;
-        for (Map.Entry<String, List<PartitionMetadata>> en : kafkaPartitions.entrySet()) {
-          String clusterId = en.getKey();
-          for (PartitionMetadata pm : en.getValue()) {
-            if (kps[i % size] == null) {
-              kps[i % size] = new HashSet<KafkaPartition>();
+        if (isInitialParitition || newWaitingPartition.size() != 0) {
+          lastRepartitionTime = System.currentTimeMillis();
+          logger.info("[ONE_TO_MANY]: Initializing partition(s)");
+          // get partition metadata for topics.
+          // Whatever operator is using high-level or simple kafka consumer, the operator always create a temporary simple kafka consumer to get the metadata of the topic
+          // The initial value of brokerList of the KafkaConsumer is used to retrieve the topic metadata
+          Map<String, List<PartitionMetadata>> kafkaPartitions = KafkaMetadataUtil.getPartitionsForTopic(getConsumer().brokers, getConsumer().getTopic());
+
+          int size = initialPartitionCount;
+          @SuppressWarnings("unchecked")
+          Set<KafkaPartition>[] kps = (Set<KafkaPartition>[])Array.newInstance((new HashSet<KafkaPartition>()).getClass(), size);
+          int i = 0;
+          for (Map.Entry<String, List<PartitionMetadata>> en : kafkaPartitions.entrySet()) {
+            String clusterId = en.getKey();
+            for (PartitionMetadata pm : en.getValue()) {
+              if (kps[i % size] == null) {
+                kps[i % size] = new HashSet<KafkaPartition>();
+              }
+              kps[i % size].add(new KafkaPartition(clusterId, consumer.topic, pm.partitionId()));
+              i++;
             }
-            kps[i % size].add(new KafkaPartition(clusterId, consumer.topic, pm.partitionId()));
-            i++;
           }
+          size = i > size ? size : i;
+          newPartitions = new ArrayList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>(size);
+          for (i = 0; i < size; i++) {
+            logger.info("[ONE_TO_MANY]: Create operator partition for kafka partition(s): {} ", StringUtils.join(kps[i], ", "));
+            newPartitions.add(createPartition(kps[i], initOffset));
+          }
+          // Add the existing partition Ids to the deleted operators
+          for (Partition<AbstractKafkaInputOperator<K>> op : partitions) {
+            deletedOperators.add(op.getPartitionedInstance().operatorId);
+          }
+          newWaitingPartition.clear();
+          resultPartitions = newPartitions;
+          numPartitionsChanged = true;
         }
-        size = i > size ? size : i;
-        newPartitions = new ArrayList<Partitioner.Partition<AbstractKafkaInputOperator<K>>>(size);
-        for (i = 0; i < size; i++) {
-          logger.info("[ONE_TO_MANY]: Create operator partition for kafka partition(s): {} ", StringUtils.join(kps[i], ", "));
-          newPartitions.add(createPartition(kps[i], initOffset));
-        }
-        // Add the existing partition Ids to the deleted operators
-        for (Partition<AbstractKafkaInputOperator<K>> op : partitions)
-        {
-          deletedOperators.add(op.getPartitionedInstance().operatorId);
-        }
-
-        newWaitingPartition.clear();
-        resultPartitions = newPartitions;
-        numPartitionsChanged = true;
-      }
-      break;
+        break;
 
-    case ONE_TO_MANY_HEURISTIC:
-      throw new UnsupportedOperationException("[ONE_TO_MANY_HEURISTIC]: Not implemented yet");
-    default:
-      break;
+      case ONE_TO_MANY_HEURISTIC:
+        throw new UnsupportedOperationException("[ONE_TO_MANY_HEURISTIC]: Not implemented yet");
+      default:
+        break;
     }
 
     if (numPartitionsChanged) {
@@ -689,7 +688,7 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
     List<KafkaConsumer.KafkaMeterStats> kmsList = new LinkedList<KafkaConsumer.KafkaMeterStats>();
     for (Stats.OperatorStats os : stats.getLastWindowedStats()) {
       if (os != null && os.counters instanceof KafkaConsumer.KafkaMeterStats) {
-        kmsList.add((KafkaConsumer.KafkaMeterStats) os.counters);
+        kmsList.add((KafkaConsumer.KafkaMeterStats)os.counters);
       }
     }
     return kmsList;
@@ -719,12 +718,12 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
       return false;
     }
 
-    if(repartitionInterval < 0){
+    if (repartitionInterval < 0) {
       // if repartition is disabled
       return false;
     }
 
-    if(t - lastRepartitionTime < repartitionInterval) {
+    if (t - lastRepartitionTime < repartitionInterval) {
       // return false if it's still within repartitionInterval since last (re)partition
       return false;
     }
@@ -747,18 +746,18 @@ public abstract class AbstractKafkaInputOperator<K extends KafkaConsumer> implem
         }
 
         Map<String, List<PartitionMetadata>> partitionsMeta = KafkaMetadataUtil.getPartitionsForTopic(consumer.brokers, consumer.getTopic());
-        if(partitionsMeta == null){
+        if (partitionsMeta == null) {
           //broker(s) has temporary issue to get metadata
           return false;
         }
         for (Map.Entry<String, List<PartitionMetadata>> en : partitionsMeta.entrySet()) {
-          if(en.getValue() == null){
+          if (en.getValue() == null) {
             //broker(s) has temporary issue to get metadata
             continue;
           }
           for (PartitionMetadata pm : en.getValue()) {
             KafkaPartition pa = new KafkaPartition(en.getKey(), consumer.topic, pm.partitionId());
-            if(!existingIds.contains(pa)){
+            if (!existingIds.contains(pa)) {
               newWaitingPartition.add(pa);
             }
           }
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java
index f0835c4..624c955 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaOutputOperator.java
@@ -16,18 +16,19 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
+
+import java.util.Properties;
+import javax.validation.constraints.NotNull;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.commons.lang3.StringUtils;
 
 import com.datatorrent.api.Context.OperatorContext;
 import com.datatorrent.api.Operator;
 import kafka.javaapi.producer.Producer;
 import kafka.producer.ProducerConfig;
-import org.apache.commons.lang3.StringUtils;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
-
-import javax.validation.constraints.NotNull;
-import java.util.Properties;
 
 /**
  * This is the base implementation of a Kafka output operator, which writes data to the Kafka message bus.
@@ -86,7 +87,8 @@ public abstract class AbstractKafkaOutputOperator<K, V> implements Operator
    * setup producer configuration.
    * @return ProducerConfig
    */
-  protected ProducerConfig createKafkaProducerConfig(){
+  protected ProducerConfig createKafkaProducerConfig()
+  {
     Properties prop = new Properties();
     for (String propString : producerProperties.split(",")) {
       if (!propString.contains("=")) {
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java
index 96dd599..f656807 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/AbstractKafkaSinglePortInputOperator.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
 
 import com.datatorrent.api.DefaultOutputPort;
 
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java
index 85cee56..5531285 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/HighlevelKafkaConsumer.java
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
 
 import java.util.HashMap;
 import java.util.List;
@@ -30,8 +30,6 @@ import java.util.concurrent.Executors;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import com.google.common.base.Joiner;
-
 import kafka.consumer.ConsumerConfig;
 import kafka.consumer.ConsumerIterator;
 import kafka.consumer.KafkaStream;
@@ -158,8 +156,8 @@ public class HighlevelKafkaConsumer extends KafkaConsumer
       Map<String, List<KafkaStream<byte[], byte[]>>> consumerMap = standardConsumer.get(e.getKey()).createMessageStreams(topicCountMap);
 
       for (final KafkaStream<byte[], byte[]> stream : consumerMap.get(topic)) {
-        consumerThreadExecutor.submit(new Runnable() {
-
+        consumerThreadExecutor.submit(new Runnable()
+        {
           KafkaPartition kp = new KafkaPartition(e.getKey(), topic, -1);
 
           public void run()
diff --git a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java
index a67ff48..f59f2c8 100644
--- a/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java
+++ b/contrib/src/main/java/org/apache/apex/malhar/contrib/kafka/KafkaConsumer.java
@@ -16,21 +16,8 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package com.datatorrent.contrib.kafka;
+package org.apache.apex.malhar.contrib.kafka;
 
-import com.datatorrent.api.Context;
-import com.esotericsoftware.kryo.serializers.FieldSerializer.Bind;
-import com.esotericsoftware.kryo.serializers.JavaSerializer;
-import com.google.common.collect.HashMultimap;
-import com.google.common.collect.Maps;
-import com.google.common.collect.SetMultimap;
-import kafka.message.Message;
-import org.apache.commons.io.IOUtils;
-import org.apache.commons.lang3.StringUtils;
-
-import javax.validation.constraints.NotNull;
-import javax.validation.constraints.Pattern;
-import javax.validation.constraints.Pattern.Flag;
 import java.io.Closeable;
 import java.io.Serializable;
 import java.util.Collection;
@@ -44,6 +31,19 @@ import java.util.concurrent.ConcurrentHashMap;
 import java.util.concurrent.Executors;
 import java.util.concurrent.ScheduledExecutorService;
 import java.util.concurrent.TimeUnit;
+import javax.validation.constraints.NotNull;
+import javax.validation.constraints.Pattern;
+import javax.validation.constraints.Pattern.Flag;
+
+import org.apache.commons.io.IOUtils;
+import org.apache.commons.lang3.StringUtils;
+import com.esotericsoftware.kryo.serializers.FieldSerializer.Bind;
+import com.esotericsoftware.kryo.serializers.JavaSerializer;
+import com.google.common.collect.HashMultimap;
+import com.google.common.collect.Maps;
+import com.google.common.collect.SetMultimap;
+import com.datatorrent.api.Context;
+import kafka.message.Message;
 
 /**
  * Base Kafka Consumer class used by kafka input operator
@@ -52,9 +52,9 @@ import java.util.concurrent.TimeUnit;
  */
 public abstract class KafkaConsumer implements Closeable
 {
-  protected final static String HIGHLEVEL_CONSUMER_ID_SUFFIX = "_stream_";
+  protected static final String HIGHLEVEL_CONSUMER_ID_SUFFIX = "_stream_";
 
-  protected final static String SIMPLE_CONSUMER_ID_SUFFIX = "_partition_";
+  protected static final String SIMPLE_CONSUMER_ID_SUFFIX = "_partition_";
   private String zookeeper;
 
   public KafkaConsumer()
@@ -111,7 +111,7 @@ public abstract class KafkaConsumer implements Closeable
    * This setting is case_insensitive
    * By default it always consume from the beginning of the queue
    */
-  @Pattern(flags={Flag.CASE_INSENSITIVE}, regexp = "earliest|latest")
+  @Pattern(flags = {Flag.CASE_INSENSITIVE}, regexp = "earliest|latest")
   protected String initialOffset = "latest";
 
 
@@ -122,17 +122,18 @@ public abstract class KafkaConsumer implements Closeable
   /**
    * This method is called in setup method of the operator
    */
-  public void create(){
+  public void create()
+  {
     initBrokers();
     holdingBuffer = new ArrayBlockingQueue<KafkaMessage>(cacheSize);
   }
 
   public void initBrokers()
   {
-    if(brokers!=null){
-      return ;
+    if (brokers != null) {
+      return;
     }
-    if(zookeeperMap !=null){
+    if (zookeeperMap != null) {
       brokers = HashMultimap.create();
       for (String clusterId: zookeeperMap.keySet()) {
         try {
@@ -158,12 +159,13 @@ public abstract class KafkaConsumer implements Closeable
   /**
    * The method is called in the deactivate method of the operator
    */
-  public void stop() {
+  public void stop()
+  {
     isAlive = false;
     statsSnapShot.stop();
     holdingBuffer.clear();
     IOUtils.closeQuietly(this);
-  };
+  }
 
   /**
    * This method is called in teardown method of the operator
@@ -227,7 +229,8 @@ public abstract class KafkaConsumer implements Closeable
   }
 
 
-  final protected void putMessage(KafkaPartition partition, Message msg, long offset) throws InterruptedException{
+  protected final void putMessage(KafkaPartition partition, Message msg, long offset) throws InterruptedException
+  {
     // block from receiving more message
     holdingBuffer.put(new KafkaMessage(partition, msg, offset));
     statsSnapShot.mark(partition, msg.payloadSize());
@@ -300,7 +303,8 @@ public abstract class KafkaConsumer implements Closeable
       totalBytesPerSec = _1minAvg[1];
     }
 
-    public void updateOffsets(Map<KafkaPartition, Long> offsets){
+    public void updateOffsets(Map<KafkaPartition, Long> offsets)
+    {
       for (Entry<KafkaPartition, Long> os : offsets.entrySet()) {
         PartitionStats ps = putPartitionStatsIfNotPresent(os.getKey());
         ps.offset = os.getValue();
@@ -325,7 +329,8 @@ public abstract class KafkaConsumer implements Closeable
       ps.brokerId = brokerId;
     }
 
-    private synchronized PartitionStats putPartitionStatsIfNotPresent(KafkaPartition kp){
+    private synchronized PartitionStats putPartitionStatsIfNotPresent(KafkaPartition kp)
+    {
       PartitionStats ps = partitionStats.get(kp);
 
       if (ps == null) {
@@ -347,6 +352,7 @@ public abstract class KafkaConsumer implements Closeable
       this.msg = msg;
       this.offSet = offset;
     }
+
     public KafkaPartition getKafkaPart()
     {
       return kafkaPart;
@@ -363,8 +369,8 @@ public abstract class KafkaConsumer implements Closeable
     }
   }
 
-  public static class KafkaMeterStatsUtil {
-
+  public static class KafkaMeterStatsUtil
+  {
     public static Map<KafkaPartition, Long> getOffsetsForPartitions(List<KafkaMeterStats> kafkaMeterStats)
     {
       Map<KafkaPartition, Long> result = Maps.newHashMap();
@@ -387,11 +393,8 @@ public abstract class KafkaConsumer implements Closeable
 
   }
 
-  public static class KafkaMeterStatsAggregator implements Context.CountersAggregator, Serializable{
-
-    /**
-     *
-     */
+  public static class KafkaMeterStatsAggregator implements Context.CountersAggregator, Serializable
+  {
     private static final long serialVersionUID = 729987800215151678L;
 
     @Override
@@ -399,7 +402,7 @@ public abstract class KafkaConsumer implements Closeable
     {
       KafkaMeterStats kms = new KafkaMeterStats();
       for (Object o : countersList) {
-        if (o instanceof KafkaMeterStats){
+        if (o instanceof KafkaMeterStats) {
           KafkaMeterStats subKMS = (KafkaMeterStats)o;
           kms.partitionStats.putAll(subKMS.partitionStats);
           kms.totalBytesPerSec += subKMS.totalBytesPerSec;
@@ -411,12 +414,8 @@ public abstract class KafkaConsumer implements Closeable
 
   }
 
-  public static class PartitionStats implements Serializable {
-
-
-    /**
-     *
-     */
+  public static class PartitionStats implements Serializable
+  {
     private static final long serialVersionUID = -6572690643487689766L;
 
     public int brokerId = -1;
@@ -431,13 +430,11 @@ public abstract class KafkaConsumer implements Closeable
 
   }
 
-
-
   /**
    * A snapshot of consuming rate within 1 min
    */
-  static class SnapShot {
-
+  static class SnapShot
+  {
     // msgs/s and bytes/s for each partition
 
     /**
@@ -485,35 +482,41 @@ public abstract class KafkaConsumer implements Closeable
     }
 
 
-    public void start(){
-      if(service==null){
+    public void start()
+    {
+      if (service == null) {
         service = Executors.newScheduledThreadPool(1);
       }
-      service.scheduleAtFixedRate(new Runnable() {
+      service.scheduleAtFixedRate(new Runnable()
+      {
         @Override
         public void run()
         {
           moveNext();
-          if(last<60)last++;
+          if (last < 60) {
+            last++;
+          }
         }
       }, 1, 1, TimeUnit.SECONDS);
 
     }
 
-    public void stop(){
-      if(service!=null){
+    public void stop()
+    {
+      if (service != null) {
         service.shutdown();
       }
     }
 
-    public synchronized void mark(KafkaPartition partition, long bytes){
+    public synchronized void mark(KafkaPartition partition, long bytes)
+    {
       msgSec[cursor]++;
       msgSec[60]++;
       bytesSec[cursor] += bytes;
       bytesSec[60] += bytes;
... 37109 lines suppressed ...

-- 
To stop receiving notification emails like this one, please contact
"commits@apex.apache.org" <co...@apex.apache.org>.