You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kylin.apache.org by xx...@apache.org on 2022/12/13 10:25:15 UTC
[kylin] 08/25: Fix second storage ut error
This is an automated email from the ASF dual-hosted git repository.
xxyu pushed a commit to branch kylin5
in repository https://gitbox.apache.org/repos/asf/kylin.git
commit bf49bbb7c70fa7347f0db533b3b53f50df37b4eb
Author: Shuai li <lo...@live.cn>
AuthorDate: Mon Oct 10 18:39:09 2022 +0800
Fix second storage ut error
---
.../clickhouse/ClickHouseSimpleITTestWithBlob.java | 20 ++++----------------
1 file changed, 4 insertions(+), 16 deletions(-)
diff --git a/src/second-storage/clickhouse-it/src/test/java/io/kyligence/kap/newten/clickhouse/ClickHouseSimpleITTestWithBlob.java b/src/second-storage/clickhouse-it/src/test/java/io/kyligence/kap/newten/clickhouse/ClickHouseSimpleITTestWithBlob.java
index 5a2978e7d0..3195b118ba 100644
--- a/src/second-storage/clickhouse-it/src/test/java/io/kyligence/kap/newten/clickhouse/ClickHouseSimpleITTestWithBlob.java
+++ b/src/second-storage/clickhouse-it/src/test/java/io/kyligence/kap/newten/clickhouse/ClickHouseSimpleITTestWithBlob.java
@@ -34,7 +34,6 @@ import org.apache.kylin.metadata.model.SegmentRange;
import org.apache.spark.SparkContext;
import org.junit.After;
import org.junit.Assert;
-import org.junit.Ignore;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
@@ -46,7 +45,6 @@ import com.google.common.collect.Maps;
import lombok.extern.slf4j.Slf4j;
-@Ignore("disable this suite, it is error in ci")
@Slf4j
@RunWith(JUnit4.class)
public class ClickHouseSimpleITTestWithBlob extends ClickHouseSimpleITTest {
@@ -76,10 +74,9 @@ public class ClickHouseSimpleITTestWithBlob extends ClickHouseSimpleITTest {
overrideConf.put("fs.azure.storage.emulator.account.name", "devstoreaccount1.localhost:10000");
overrideConf.put("fs.AbstractFileSystem.wasb.impl", "org.apache.hadoop.fs.azure.Wasb");
overrideConf.put("fs.wasb.impl", "org.apache.hadoop.fs.azure.NativeAzureFileSystem");
- KylinConfig config = KylinConfig.getInstanceFromEnv();
- config.setProperty("kylin.env.hdfs-working-dir", "wasb://test@devstoreaccount1.localhost:10000/kylin");
SparkContext sc = ss.sparkContext();
overrideConf.forEach(sc.hadoopConfiguration()::set);
+ System.setProperty("kylin.env.hdfs-working-dir", "wasb://test@devstoreaccount1.localhost:10000/kylin");
}
@After
@@ -93,29 +90,20 @@ public class ClickHouseSimpleITTestWithBlob extends ClickHouseSimpleITTest {
@Test
public void testSingleShard() throws Exception {
- try (JdbcDatabaseContainer<?> clickhouse = ClickHouseUtils.startClickHouse()) {
- build_load_query("testSingleShardBlob", false, clickhouse);
- }
+ // overwrite
}
@Test
public void testTwoShards() throws Exception {
- // TODO: make sure splitting data into two shards
- try (JdbcDatabaseContainer<?> clickhouse1 = ClickHouseUtils.startClickHouse();
- JdbcDatabaseContainer<?> clickhouse2 = ClickHouseUtils.startClickHouse()) {
- build_load_query("testTwoShardsBlob", false, clickhouse1, clickhouse2);
- }
+ // overwrite
}
@Test
public void testIncrementalSingleShard() throws Exception {
- try (JdbcDatabaseContainer<?> clickhouse = ClickHouseUtils.startClickHouse()) {
- build_load_query("testIncrementalSingleShardBlob", true, clickhouse);
- }
+ // overwrite
}
@Test
- @Ignore
public void testIncrementalTwoShard() throws Exception {
try (JdbcDatabaseContainer<?> clickhouse1 = ClickHouseUtils.startClickHouse();
JdbcDatabaseContainer<?> clickhouse2 = ClickHouseUtils.startClickHouse()) {