You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@bahir.apache.org by lr...@apache.org on 2019/05/19 21:08:09 UTC
[bahir-flink] branch master updated: [BAHIR-206] Update Flink to
release 1.8.0
This is an automated email from the ASF dual-hosted git repository.
lresende pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/bahir-flink.git
The following commit(s) were added to refs/heads/master by this push:
new 568bb5b [BAHIR-206] Update Flink to release 1.8.0
568bb5b is described below
commit 568bb5bcfea35e7ba258e0c8d2e12388e0dfe868
Author: eskabetxe <bo...@boto.pro>
AuthorDate: Thu May 16 15:57:17 2019 +0200
[BAHIR-206] Update Flink to release 1.8.0
Closes #54
---
.travis.yml | 6 +++---
flink-connector-activemq/pom.xml | 7 ++++++
.../activemq/ActiveMQConnectorITCase.java | 25 ++++------------------
flink-connector-netty/pom.xml | 3 +--
.../netty/example/StreamSqlExample.scala | 5 ++---
pom.xml | 2 +-
6 files changed, 18 insertions(+), 30 deletions(-)
diff --git a/.travis.yml b/.travis.yml
index 191591d..d876e7d 100644
--- a/.travis.yml
+++ b/.travis.yml
@@ -37,13 +37,13 @@ jdk:
env:
- |
- FLINK_VERSION="1.7.2" SCALA_VERSION="2.11" DOCKER="false"
+ FLINK_VERSION="1.8.0" SCALA_VERSION="2.11" DOCKER="false"
PROJECTS="flink-connector-activemq,flink-connector-akka,flink-connector-influxdb,flink-connector-netty,flink-connector-redis,flink-library-siddhi"
- |
- FLINK_VERSION="1.7.2" SCALA_VERSION="2.11" DOCKER="true"
+ FLINK_VERSION="1.8.0" SCALA_VERSION="2.11" DOCKER="true"
PROJECTS="flink-connector-flume"
- |
- FLINK_VERSION="1.7.2" SCALA_VERSION="2.11" DOCKER="true"
+ FLINK_VERSION="1.8.0" SCALA_VERSION="2.11" DOCKER="true"
PROJECTS="flink-connector-kudu"
before_install:
diff --git a/flink-connector-activemq/pom.xml b/flink-connector-activemq/pom.xml
index 8efbedb..8f7cdaf 100644
--- a/flink-connector-activemq/pom.xml
+++ b/flink-connector-activemq/pom.xml
@@ -78,6 +78,13 @@ under the License.
<dependency>
<groupId>org.apache.flink</groupId>
+ <artifactId>flink-test-utils_${scala.binary.version}</artifactId>
+ <version>${flink.version}</version>
+ <scope>test</scope>
+ </dependency>
+
+ <dependency>
+ <groupId>org.apache.flink</groupId>
<artifactId>flink-runtime_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
<type>test-jar</type>
diff --git a/flink-connector-activemq/src/test/java/org/apache/flink/streaming/connectors/activemq/ActiveMQConnectorITCase.java b/flink-connector-activemq/src/test/java/org/apache/flink/streaming/connectors/activemq/ActiveMQConnectorITCase.java
index 1160128..6e3a247 100644
--- a/flink-connector-activemq/src/test/java/org/apache/flink/streaming/connectors/activemq/ActiveMQConnectorITCase.java
+++ b/flink-connector-activemq/src/test/java/org/apache/flink/streaming/connectors/activemq/ActiveMQConnectorITCase.java
@@ -24,7 +24,6 @@ import org.apache.flink.api.common.serialization.SimpleStringSchema;
import org.apache.flink.configuration.ConfigConstants;
import org.apache.flink.configuration.Configuration;
import org.apache.flink.configuration.TaskManagerOptions;
-import org.apache.flink.runtime.minicluster.LocalFlinkMiniCluster;
import org.apache.flink.runtime.state.FunctionSnapshotContext;
import org.apache.flink.streaming.api.datastream.DataStreamSource;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
@@ -33,10 +32,8 @@ import org.apache.flink.streaming.api.functions.source.SourceFunction;
import org.apache.flink.streaming.api.operators.StreamingRuntimeContext;
import org.apache.flink.streaming.api.watermark.Watermark;
import org.apache.flink.test.util.SuccessException;
-import org.junit.AfterClass;
import org.junit.BeforeClass;
import org.junit.Test;
-import org.junit.jupiter.api.Disabled;
import scala.concurrent.duration.Deadline;
import scala.concurrent.duration.FiniteDuration;
@@ -48,37 +45,23 @@ import static org.junit.Assert.assertEquals;
import static org.mockito.Mockito.mock;
import static org.mockito.Mockito.when;
-
-@Disabled("failing")
public class ActiveMQConnectorITCase {
public static final int MESSAGES_NUM = 10000;
public static final String QUEUE_NAME = "queue";
public static final String TOPIC_NAME = "topic";
- private static LocalFlinkMiniCluster flink;
- private static int flinkPort;
+
+ private static Configuration flinkConfig;
@BeforeClass
public static void beforeClass() {
// start also a re-usable Flink mini cluster
- Configuration flinkConfig = new Configuration();
+ flinkConfig = new Configuration();
flinkConfig.setInteger(ConfigConstants.LOCAL_NUMBER_TASK_MANAGER, 1);
flinkConfig.setInteger(TaskManagerOptions.NUM_TASK_SLOTS.key(), 8);
flinkConfig.setInteger(TaskManagerOptions.MANAGED_MEMORY_SIZE.key(), 16);
flinkConfig.setString(ConfigConstants.RESTART_STRATEGY_FIXED_DELAY_DELAY, "0 s");
- flink = new LocalFlinkMiniCluster(flinkConfig, false);
- flink.start();
-
- flinkPort = flink.getLeaderRPCPort();
- }
-
- @AfterClass
- public static void afterClass() {
- flinkPort = -1;
- if (flink != null) {
- flink.startInternalShutdown();
- }
}
@Test
@@ -126,7 +109,7 @@ public class ActiveMQConnectorITCase {
}
private StreamExecutionEnvironment createExecutionEnvironment() {
- StreamExecutionEnvironment env = StreamExecutionEnvironment.createRemoteEnvironment("localhost", flinkPort);
+ StreamExecutionEnvironment env = StreamExecutionEnvironment.createLocalEnvironmentWithWebUI(flinkConfig);
env.setRestartStrategy(RestartStrategies.noRestart());
env.getConfig().disableSysoutLogging();
return env;
diff --git a/flink-connector-netty/pom.xml b/flink-connector-netty/pom.xml
index 7a9097a..634b24f 100644
--- a/flink-connector-netty/pom.xml
+++ b/flink-connector-netty/pom.xml
@@ -68,9 +68,8 @@
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
- <artifactId>flink-table_${scala.binary.version}</artifactId>
+ <artifactId>flink-table-planner_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
- <scope>test</scope>
</dependency>
</dependencies>
diff --git a/flink-connector-netty/src/test/scala/org/apache/flink/streaming/connectors/netty/example/StreamSqlExample.scala b/flink-connector-netty/src/test/scala/org/apache/flink/streaming/connectors/netty/example/StreamSqlExample.scala
index 5167b3e..9f4d0be 100644
--- a/flink-connector-netty/src/test/scala/org/apache/flink/streaming/connectors/netty/example/StreamSqlExample.scala
+++ b/flink-connector-netty/src/test/scala/org/apache/flink/streaming/connectors/netty/example/StreamSqlExample.scala
@@ -18,9 +18,8 @@ package org.apache.flink.streaming.connectors.netty.example
import org.apache.flink.api.java.utils.ParameterTool
import org.apache.flink.api.scala._
-import org.apache.flink.table.api.scala._
-import org.apache.flink.table.api.TableEnvironment
import org.apache.flink.streaming.api.scala.{DataStream, StreamExecutionEnvironment}
+import org.apache.flink.table.api.scala._
/**
* Simple example for demonstrating the use of SQL on a Stream Table.
@@ -41,7 +40,7 @@ object StreamSqlExample {
// set up execution environment
val env = StreamExecutionEnvironment.getExecutionEnvironment
- val tEnv = TableEnvironment.getTableEnvironment(env)
+ val tEnv = StreamTableEnvironment.create(env)
val spec = if (param.get("tcp") == "true") {
new TcpReceiverSource(7070, Some("http://localhost:9090/cb"))
diff --git a/pom.xml b/pom.xml
index 8c5e717..2706d50 100644
--- a/pom.xml
+++ b/pom.xml
@@ -96,7 +96,7 @@
<log4j.version>1.2.17</log4j.version>
<!-- Flink version -->
- <flink.version>1.7.2</flink.version>
+ <flink.version>1.8.0</flink.version>
<junit.jupiter.version>5.3.1</junit.jupiter.version>
<junit.groups></junit.groups>