You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by zentol <gi...@git.apache.org> on 2017/05/25 22:18:58 UTC

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

GitHub user zentol opened a pull request:

    https://github.com/apache/flink/pull/3992

    [FLINK-6711] Activate strict checkstyle for flink-connectors

    

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/zentol/flink 6711_cs_con

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/flink/pull/3992.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #3992
    
----
commit b9c4449d56cf282e58d2dc325286397a4f3348b6
Author: zentol <ch...@apache.org>
Date:   2017-05-24T20:29:09Z

    [FLINK-6711] Activate strict checkstyle for flink-connector-twitter

commit 6fa5187353e018a7f33a002021f418b9910107c1
Author: zentol <ch...@apache.org>
Date:   2017-05-24T20:31:36Z

    [FLINK-6711] Activate strict checkstyle for flink-connector-nifi

commit 6a2aa3f1e37fc89ef698adbae136fb569d311042
Author: zentol <ch...@apache.org>
Date:   2017-05-24T20:49:55Z

    [FLINK-6711] Activate strict checkstyle for flink-connector-rabbitmq

commit 7e284fbfa9504d225ca5e395d61f0e2dab0aa3fc
Author: zentol <ch...@apache.org>
Date:   2017-05-24T20:55:11Z

    [FLINK-6711] Activate strict checkstyle for flink-jdbc

commit e2f1d7b37fe41fbd5c9c7d802698010eafac59a7
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:12:36Z

    [FLINK-6711] Activate strict checkstyle for flink-hbase

commit cc39da15608baa664fa5b52813ddf0e095e9da14
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:14:27Z

    [FLINK-6711] Activate strict checkstyle for flink-hcatalog

commit d135e45fbf032b3f3978c11b738887326d7e7b03
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:20:29Z

    [FLINK-6711] Activate strict checkstyle for flink-elasticsearch*

commit fb2e7308c78501ac7cfaa898b698786a2e3da0a1
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:55:15Z

    [FLINK-6711] Activate strict checkstyle for flink-connector-kafka*

commit 9e8ced0abf1d7b07a741912464f2fcc6e1f67c84
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:56:16Z

    [FLINK-6711] Activate strict checkstyle for flink-connector-kinesis

commit 5c66d3406bd078c43c7bf3af5e36bd4fbb7321c0
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:56:53Z

    [FLINK-6711] Activate strict checkstyle for flink-hadoop-compatibility

commit 2f38a2ffa07ad79da4b49127501cb76afb4a6dd9
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:57:18Z

    [FLINK-6711] Activate strict checkstyle for flink-connector-filesystem

commit bbc1e36331afab50d3b831cc0008461d23206604
Author: zentol <ch...@apache.org>
Date:   2017-05-24T21:57:46Z

    [FLINK-6711] Activate strict checkstyle for flink-connector-cassandra

commit c7376088b0695f26cc09ebac957cd393f390dcbf
Author: zentol <ch...@apache.org>
Date:   2017-05-25T22:15:34Z

    [FLINK-6711] Activate strict checkstyle for flink-avro

commit e982d89476b5f0b5a4610fca7949df8390e124f5
Author: zentol <ch...@apache.org>
Date:   2017-05-25T16:52:40Z

    [FLINK-6711] Activate strict checkstyle for flink-connectors

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118727020
  
    --- Diff: flink-connectors/flink-jdbc/src/main/java/org/apache/flink/api/java/io/jdbc/split/GenericParameterValuesProvider.java ---
    @@ -15,22 +15,22 @@
      * See the License for the specific language governing permissions and
      * limitations under the License.
      */
    -package org.apache.flink.api.java.io.jdbc.split;
     
    -import java.io.Serializable;
    +package org.apache.flink.api.java.io.jdbc.split;
     
     import org.apache.flink.api.java.io.jdbc.JDBCInputFormat;
     
    -/** 
    - * 
    +import java.io.Serializable;
    +
    +/**
      * This splits generator actually does nothing but wrapping the query parameters
      * computed by the user before creating the {@link JDBCInputFormat} instance.
    - * 
    + *
      * */
    --- End diff --
    
    Empty line and extra asterisk.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118739552
  
    --- Diff: flink-connectors/flink-avro/src/main/java/org/apache/flink/api/avro/FSDataInputStreamWrapper.java ---
    @@ -16,20 +16,19 @@
      * limitations under the License.
      */
     
    -
     package org.apache.flink.api.avro;
     
    -import java.io.Closeable;
    -import java.io.IOException;
    +import org.apache.flink.core.fs.FSDataInputStream;
     
     import org.apache.avro.file.SeekableInput;
    -import org.apache.flink.core.fs.FSDataInputStream;
     
    +import java.io.Closeable;
    +import java.io.IOException;
     
     /**
    - * Code copy pasted from org.apache.avro.mapred.FSInput (which is Apache licensed as well)
    - * 
    - * The wrapper keeps track of the position in the data stream.
    + * Code copy pasted from org.apache.avro.mapred.FSInput (which is Apache licensed as well.
    --- End diff --
    
    Lost the closing parenthesis.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118734813
  
    --- Diff: flink-connectors/flink-connector-kafka-base/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTableSourceTestBase.java ---
    @@ -47,10 +53,14 @@
     		BasicTypeInfo.LONG_TYPE_INFO };
     	private static final Properties PROPERTIES = createSourceProperties();
     
    -	// Avro record that matches above schema
    +	/**
    +	 * Avro record that matches above schema.
    +	 */
     	public static class AvroSpecificRecord extends SpecificRecordBase {
     
    +		//CHECKSTYLE.OFF: ConstantNameCheck|
    --- End diff --
    
    Is this intended to have a trailing pipe?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118728159
  
    --- Diff: flink-connectors/flink-hbase/src/test/java/org/apache/flink/addons/hbase/example/HBaseWriteStreamExample.java ---
    @@ -17,27 +17,27 @@
     
     package org.apache.flink.addons.hbase.example;
     
    -import java.io.IOException;
    -
     import org.apache.flink.api.common.io.OutputFormat;
     import org.apache.flink.configuration.Configuration;
     import org.apache.flink.streaming.api.datastream.DataStream;
     import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
     import org.apache.flink.streaming.api.functions.source.SourceFunction;
    +
     import org.apache.hadoop.hbase.HBaseConfiguration;
     import org.apache.hadoop.hbase.client.HTable;
     import org.apache.hadoop.hbase.client.Put;
     import org.apache.hadoop.hbase.util.Bytes;
     
    +import java.io.IOException;
    +
     /**
    - * 
      * This is an example how to write streams into HBase. In this example the
      * stream will be written into a local Hbase but it is possible to adapt this
      * example for an HBase running in a cloud. You need a running local HBase with a
      * table "flinkExample" and a column "entry". If your HBase configuration does
      * not fit the hbase-site.xml in the resource folder then you gave to delete temporary this
      * hbase-site.xml to execute the example properly.
    - * 
    + *
    --- End diff --
    
    Empty line.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118726185
  
    --- Diff: flink-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQSource.java ---
    @@ -105,7 +105,7 @@ public RMQSource(RMQConnectionConfig rmqConnectionConfig, String queueName,
     	/**
     	 * Creates a new RabbitMQ source. For exactly-once, you must set the correlation ids of messages
     	 * at the producer. The correlation id must be unique. Otherwise the behavior of the source is
    -	 * undefined. In doubt, set {@param usesCorrelationId} to false. When correlation ids are not
    +	 * undefined. In doubt, set usesCorrelationId to false. When correlation ids are not
    --- End diff --
    
    "If in doubt"?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by zentol <gi...@git.apache.org>.
Github user zentol commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118736979
  
    --- Diff: flink-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/Kafka08JsonTableSink.java ---
    @@ -42,7 +43,7 @@ public Kafka08JsonTableSink(String topic, Properties properties, FlinkKafkaParti
     	}
     
     	/**
    -	 * Creates {@link KafkaTableSink} for Kafka 0.8
    +	 * Creates {@link KafkaTableSink} for Kafka 0.8 .
    --- End diff --
    
    This is somewhat intended, to clarify that it is the end of the sentence and not an incomplete version. Don't mind removing the space though.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink issue #3992: [FLINK-6711] Activate strict checkstyle for flink-connect...

Posted by zentol <gi...@git.apache.org>.
Github user zentol commented on the issue:

    https://github.com/apache/flink/pull/3992
  
    @greghogan I've fixed all the issues you found. I don't know what to do about the checkstyle disabling though using `//CHECKSTYLE.OFF: ConstantNameCheck|`. The regex evaluation appears to be broken, but maybe it's only me?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118738325
  
    --- Diff: flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/RollingSink.java ---
    @@ -266,7 +265,7 @@
     
     	/**
     	 * We use reflection to get the .truncate() method, this is only available starting with
    -	 * Hadoop 2.7
    +	 * Hadoop 2.7 .
    --- End diff --
    
    Another space before period.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by asfgit <gi...@git.apache.org>.
Github user asfgit closed the pull request at:

    https://github.com/apache/flink/pull/3992


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118725529
  
    --- Diff: flink-connectors/flink-connector-nifi/src/main/java/org/apache/flink/streaming/connectors/nifi/NiFiDataPacket.java ---
    @@ -14,13 +14,14 @@
      * See the License for the specific language governing permissions and
      * limitations under the License.
      */
    +
     package org.apache.flink.streaming.connectors.nifi;
     
     import java.util.Map;
     
     /**
    - * <p>
    - * The NiFiDataPacket provides a packaging around a NiFi FlowFile. It wraps both
    + *
    + * <p>The NiFiDataPacket provides a packaging around a NiFi FlowFile. It wraps both
    --- End diff --
    
    Remove opening and closing paragraph tags and blank line.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118732112
  
    --- Diff: flink-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/Kafka08JsonTableSink.java ---
    @@ -15,23 +15,24 @@
      * See the License for the specific language governing permissions and
      * limitations under the License.
      */
    +
     package org.apache.flink.streaming.connectors.kafka;
     
     import org.apache.flink.streaming.connectors.kafka.partitioner.FlinkKafkaDelegatePartitioner;
     import org.apache.flink.streaming.connectors.kafka.partitioner.FlinkKafkaPartitioner;
    -import org.apache.flink.types.Row;
     import org.apache.flink.streaming.connectors.kafka.partitioner.KafkaPartitioner;
     import org.apache.flink.streaming.util.serialization.SerializationSchema;
    +import org.apache.flink.types.Row;
     
     import java.util.Properties;
     
     /**
      * Kafka 0.8 {@link KafkaTableSink} that serializes data in JSON format.
      */
     public class Kafka08JsonTableSink extends KafkaJsonTableSink {
    -	
    +
     	/**
    -	 * Creates {@link KafkaTableSink} for Kafka 0.8
    +	 * Creates {@link KafkaTableSink} for Kafka 0.8 .
    --- End diff --
    
    Extra space before period.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by zentol <gi...@git.apache.org>.
Github user zentol commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118736755
  
    --- Diff: flink-connectors/flink-connector-kafka-base/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTableSourceTestBase.java ---
    @@ -47,10 +53,14 @@
     		BasicTypeInfo.LONG_TYPE_INFO };
     	private static final Properties PROPERTIES = createSourceProperties();
     
    -	// Avro record that matches above schema
    +	/**
    +	 * Avro record that matches above schema.
    +	 */
     	public static class AvroSpecificRecord extends SpecificRecordBase {
     
    +		//CHECKSTYLE.OFF: ConstantNameCheck|
    --- End diff --
    
    For some reason checkstyle kept failing without the pipe; i have no idea why.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118738883
  
    --- Diff: flink-connectors/flink-connector-filesystem/src/test/java/org/apache/flink/streaming/connectors/fs/RollingSinkSecuredITCase.java ---
    @@ -47,21 +48,21 @@
     import java.util.HashMap;
     import java.util.Map;
     
    +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY;
    +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_ADDRESS_KEY;
    +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_HOST_NAME_KEY;
     import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_HTTP_ADDRESS_KEY;
    +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_KEYTAB_FILE_KEY;
    +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_USER_NAME_KEY;
    +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_ENCRYPT_DATA_TRANSFER_KEY;
    +import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_HTTP_POLICY_KEY;
     import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_KEYTAB_FILE_KEY;
     import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NAMENODE_USER_NAME_KEY;
    -import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_USER_NAME_KEY;
    -import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_KEYTAB_FILE_KEY;
     import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_WEB_AUTHENTICATION_KERBEROS_PRINCIPAL_KEY;
    -import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY;
    -import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_HTTP_POLICY_KEY;
    -import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_ENCRYPT_DATA_TRANSFER_KEY;
    -import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_ADDRESS_KEY;
    -import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_DATANODE_HOST_NAME_KEY;
     
     /**
      * Tests for running {@link RollingSinkSecuredITCase} which is an extension of {@link RollingSink} in secure environment
    - * Note: only executed for Hadoop version > 3.x.x
    + * Note: only executed for Hadoop version > 3.x.x .
    --- End diff --
    
    And another.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118732131
  
    --- Diff: flink-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/Kafka08JsonTableSink.java ---
    @@ -42,7 +43,7 @@ public Kafka08JsonTableSink(String topic, Properties properties, FlinkKafkaParti
     	}
     
     	/**
    -	 * Creates {@link KafkaTableSink} for Kafka 0.8
    +	 * Creates {@link KafkaTableSink} for Kafka 0.8 .
    --- End diff --
    
    Extra space before period.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118738236
  
    --- Diff: flink-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/Kafka08JsonTableSink.java ---
    @@ -42,7 +43,7 @@ public Kafka08JsonTableSink(String topic, Properties properties, FlinkKafkaParti
     	}
     
     	/**
    -	 * Creates {@link KafkaTableSink} for Kafka 0.8
    +	 * Creates {@link KafkaTableSink} for Kafka 0.8 .
    --- End diff --
    
    Okay, it was clearly a pattern but wasn't sure if it was an unexpected programmatic replacement. I think it's find leaving as is but expect that someone will eventually flag it for fixing.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by zentol <gi...@git.apache.org>.
Github user zentol commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118797294
  
    --- Diff: flink-connectors/flink-connector-kafka-base/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaTableSourceTestBase.java ---
    @@ -47,10 +53,14 @@
     		BasicTypeInfo.LONG_TYPE_INFO };
     	private static final Properties PROPERTIES = createSourceProperties();
     
    -	// Avro record that matches above schema
    +	/**
    +	 * Avro record that matches above schema.
    +	 */
     	public static class AvroSpecificRecord extends SpecificRecordBase {
     
    +		//CHECKSTYLE.OFF: ConstantNameCheck|
    --- End diff --
    
    Figured this out, i was supplying the wrong rule to disable, should be `StaticVariableNameCheck` instead.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118731513
  
    --- Diff: flink-connectors/flink-connector-kafka-0.8/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumer08.java ---
    @@ -102,21 +99,21 @@
     
     	private static final long serialVersionUID = -6272159445203409112L;
     
    -	/** Configuration key for the number of retries for getting the partition info */
    +	/** Configuration key for the number of retries for getting the partition info. */
     	public static final String GET_PARTITIONS_RETRIES_KEY = "flink.get-partitions.retry";
     
     	/** Default number of retries for getting the partition info. One retry means going through the full list of brokers */
     	public static final int DEFAULT_GET_PARTITIONS_RETRIES = 3;
     
     	// ------------------------------------------------------------------------
     
    -	/** The properties to parametrize the Kafka consumer and ZooKeeper client */ 
    +	/** The properties to parametrize the Kafka consumer and ZooKeeper client. */
     	private final Properties kafkaProperties;
     
     	// ------------------------------------------------------------------------
     
     	/**
    -	 * Creates a new Kafka streaming source consumer for Kafka 0.8.x
    +	 * Creates a new Kafka streaming source consumer for Kafka 0.8.x .
    --- End diff --
    
    Also has an extra space.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118729203
  
    --- Diff: flink-connectors/flink-connector-elasticsearch-base/src/main/java/org/apache/flink/streaming/connectors/elasticsearch/ElasticsearchSinkBase.java ---
    @@ -80,11 +79,21 @@
     	public static final String CONFIG_KEY_BULK_FLUSH_BACKOFF_RETRIES = "bulk.flush.backoff.retries";
     	public static final String CONFIG_KEY_BULK_FLUSH_BACKOFF_DELAY = "bulk.flush.backoff.delay";
     
    +	/**
    +	 * Used to control whether the retry delay should increase exponentially or remain constant.
    +	 */
     	public enum FlushBackoffType {
     		CONSTANT,
     		EXPONENTIAL
     	}
     
    +	/**
    +	 * Provides a backoff policy for bulk requests. Whenever a bulk request is rejected due to resource constraints
    +	 * (i.e. the client's internal thread pool is full), the backoff policy decides how long the bulk processor will
    +	 * wait before the operation is retried internally.
    +	 *
    +	 * <p>This ia a proxy for version specific backoff policies.
    --- End diff --
    
    "ia" => "is"


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118730394
  
    --- Diff: flink-connectors/flink-connector-kafka-0.10/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumer010.java ---
    @@ -62,11 +62,10 @@
     
     	private static final long serialVersionUID = 2324564345203409112L;
     
    -
     	// ------------------------------------------------------------------------
     
     	/**
    -	 * Creates a new Kafka streaming source consumer for Kafka 0.10.x
    +	 * Creates a new Kafka streaming source consumer for Kafka 0.10.x .
    --- End diff --
    
    Extra space before period.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink issue #3992: [FLINK-6711] Activate strict checkstyle for flink-connect...

Posted by zentol <gi...@git.apache.org>.
Github user zentol commented on the issue:

    https://github.com/apache/flink/pull/3992
  
    merging.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink pull request #3992: [FLINK-6711] Activate strict checkstyle for flink-...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on a diff in the pull request:

    https://github.com/apache/flink/pull/3992#discussion_r118735031
  
    --- Diff: flink-connectors/flink-connector-kafka-base/src/test/java/org/apache/flink/streaming/connectors/kafka/testutils/FailingIdentityMapper.java ---
    @@ -22,30 +22,34 @@
     import org.apache.flink.configuration.Configuration;
     import org.apache.flink.runtime.state.CheckpointListener;
     import org.apache.flink.streaming.api.checkpoint.ListCheckpointed;
    +
     import org.slf4j.Logger;
     import org.slf4j.LoggerFactory;
     
     import java.util.Collections;
     import java.util.List;
     
    -
    -public class FailingIdentityMapper<T> extends RichMapFunction<T,T> implements
    +/**
    + * A {@link RichMapFunction} that fails after the configured number of records have been processed.
    + * @param <T>
    --- End diff --
    
    Add extra line.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

[GitHub] flink issue #3992: [FLINK-6711] Activate strict checkstyle for flink-connect...

Posted by greghogan <gi...@git.apache.org>.
Github user greghogan commented on the issue:

    https://github.com/apache/flink/pull/3992
  
    +1


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---