You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@flink.apache.org by lz...@apache.org on 2020/06/03 02:39:56 UTC
[flink] branch release-1.11 updated: [FLINK-17937][hive] Change
some hive connector tests to IT cases
This is an automated email from the ASF dual-hosted git repository.
lzljs3620320 pushed a commit to branch release-1.11
in repository https://gitbox.apache.org/repos/asf/flink.git
The following commit(s) were added to refs/heads/release-1.11 by this push:
new 636901c [FLINK-17937][hive] Change some hive connector tests to IT cases
636901c is described below
commit 636901c1e29b63378faf9dac85465f5ed0481649
Author: Rui Li <li...@apache.org>
AuthorDate: Wed Jun 3 10:38:29 2020 +0800
[FLINK-17937][hive] Change some hive connector tests to IT cases
This closes #12333
---
.../{HiveDialectTest.java => HiveDialectITCase.java} | 2 +-
...LookupJoinTest.java => HiveLookupJoinITCase.java} | 20 +++++---------------
...veTableSinkTest.java => HiveTableSinkITCase.java} | 2 +-
...bleSourceTest.java => HiveTableSourceITCase.java} | 4 ++--
...torTest.java => TableEnvHiveConnectorITCase.java} | 2 +-
5 files changed, 10 insertions(+), 20 deletions(-)
diff --git a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveDialectTest.java b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveDialectITCase.java
similarity index 99%
rename from flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveDialectTest.java
rename to flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveDialectITCase.java
index 9b6df99..12730b5 100644
--- a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveDialectTest.java
+++ b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveDialectITCase.java
@@ -71,7 +71,7 @@ import static org.junit.Assert.assertTrue;
/**
* Test Hive syntax when Hive dialect is used.
*/
-public class HiveDialectTest {
+public class HiveDialectITCase {
private TableEnvironment tableEnv;
private HiveCatalog hiveCatalog;
diff --git a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinTest.java b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinITCase.java
similarity index 87%
rename from flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinTest.java
rename to flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinITCase.java
index 88fae01..90490d7 100644
--- a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinTest.java
+++ b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinITCase.java
@@ -30,16 +30,14 @@ import org.apache.flink.table.factories.TableSourceFactoryContextImpl;
import org.apache.flink.table.filesystem.FileSystemLookupFunction;
import org.apache.flink.table.filesystem.FileSystemOptions;
import org.apache.flink.table.planner.factories.utils.TestCollectionTableFactory;
+import org.apache.flink.table.planner.runtime.utils.TableEnvUtil;
import org.apache.flink.types.Row;
import org.apache.flink.shaded.guava18.com.google.common.collect.Lists;
-import com.klarna.hiverunner.HiveShell;
-import com.klarna.hiverunner.annotations.HiveSQL;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
-import org.junit.runner.RunWith;
import java.time.Duration;
import java.util.Arrays;
@@ -50,8 +48,7 @@ import static org.junit.Assert.assertEquals;
/**
* Test lookup join of hive tables.
*/
-@RunWith(FlinkStandaloneHiveRunner.class)
-public class HiveLookupJoinTest {
+public class HiveLookupJoinITCase {
private TableEnvironment tableEnv;
private HiveCatalog hiveCatalog;
@@ -60,7 +57,7 @@ public class HiveLookupJoinTest {
public void setup() {
EnvironmentSettings settings = EnvironmentSettings.newInstance().inStreamingMode().useBlinkPlanner().build();
tableEnv = TableEnvironment.create(settings);
- hiveCatalog = HiveTestUtils.createHiveCatalog(hiveShell.getHiveConf());
+ hiveCatalog = HiveTestUtils.createHiveCatalog();
tableEnv.registerCatalog(hiveCatalog.getName(), hiveCatalog);
tableEnv.useCatalog(hiveCatalog.getName());
}
@@ -72,9 +69,6 @@ public class HiveLookupJoinTest {
}
}
- @HiveSQL(files = {})
- private static HiveShell hiveShell;
-
@Test
public void test() throws Exception {
// create the hive build table
@@ -92,12 +86,8 @@ public class HiveLookupJoinTest {
assertEquals(Duration.ofMinutes(5), lookupFunction.getCacheTTL());
try {
- HiveTestUtils.createTextTableInserter(hiveShell, "default", "build")
- .addRow(new Object[]{1, "a", 10})
- .addRow(new Object[]{2, "a", 21})
- .addRow(new Object[]{2, "b", 22})
- .addRow(new Object[]{3, "c", 33})
- .commit();
+ TableEnvUtil.execInsertSqlAndWaitResult(tableEnv,
+ "insert into build values (1,'a',10),(2,'a',21),(2,'b',22),(3,'c',33)");
TestCollectionTableFactory.initData(
Arrays.asList(Row.of(1, "a"), Row.of(1, "c"), Row.of(2, "b"), Row.of(2, "c"), Row.of(3, "c"), Row.of(4, "d")));
diff --git a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSinkTest.java b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSinkITCase.java
similarity index 99%
rename from flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSinkTest.java
rename to flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSinkITCase.java
index 2cc03de..a73bce9 100644
--- a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSinkTest.java
+++ b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSinkITCase.java
@@ -78,7 +78,7 @@ import static org.junit.Assert.assertEquals;
* Tests {@link HiveTableSink}.
*/
@RunWith(FlinkStandaloneHiveRunner.class)
-public class HiveTableSinkTest {
+public class HiveTableSinkITCase {
@HiveSQL(files = {})
private static HiveShell hiveShell;
diff --git a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSourceTest.java b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSourceITCase.java
similarity index 99%
rename from flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSourceTest.java
rename to flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSourceITCase.java
index f36a60a..317943a 100644
--- a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSourceTest.java
+++ b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableSourceITCase.java
@@ -99,7 +99,7 @@ import static org.mockito.Mockito.spy;
* Tests {@link HiveTableSource}.
*/
@RunWith(FlinkStandaloneHiveRunner.class)
-public class HiveTableSourceTest extends BatchAbstractTestBase {
+public class HiveTableSourceITCase extends BatchAbstractTestBase {
@HiveSQL(files = {})
private static HiveShell hiveShell;
@@ -561,7 +561,7 @@ public class HiveTableSourceTest extends BatchAbstractTestBase {
Runnable runnable = () -> {
for (int i = 0; i < 3; ++i) {
- hiveShell.execute("insert into source_db." + tblName + " values (1,'a'), (2,'b')");
+ hiveShell.execute("insert into table source_db." + tblName + " values (1,'a'), (2,'b')");
try {
Thread.sleep(2_000);
} catch (InterruptedException e) {
diff --git a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/TableEnvHiveConnectorTest.java b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/TableEnvHiveConnectorITCase.java
similarity index 99%
rename from flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/TableEnvHiveConnectorTest.java
rename to flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/TableEnvHiveConnectorITCase.java
index 2dcd21b..2da7d10 100644
--- a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/TableEnvHiveConnectorTest.java
+++ b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/TableEnvHiveConnectorITCase.java
@@ -73,7 +73,7 @@ import static org.junit.Assert.assertTrue;
* Test hive connector with table API.
*/
@RunWith(FlinkStandaloneHiveRunner.class)
-public class TableEnvHiveConnectorTest {
+public class TableEnvHiveConnectorITCase {
@HiveSQL(files = {})
private static HiveShell hiveShell;