You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kyuubi.apache.org by ch...@apache.org on 2023/02/19 08:31:40 UTC
[kyuubi] branch master updated: [KYUUBI #4338][FOLLOWUP] Fix K8s integration tests
This is an automated email from the ASF dual-hosted git repository.
chengpan pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/kyuubi.git
The following commit(s) were added to refs/heads/master by this push:
new a896e95bd [KYUUBI #4338][FOLLOWUP] Fix K8s integration tests
a896e95bd is described below
commit a896e95bd4645244a6ee03ee4d928bdc14633b1e
Author: Cheng Pan <ch...@apache.org>
AuthorDate: Sun Feb 19 16:31:26 2023 +0800
[KYUUBI #4338][FOLLOWUP] Fix K8s integration tests
### _Why are the changes needed?_
Correct the Spark image tag to recover "Kyuubi Server On Kubernetes Integration Test"
### _How was this patch tested?_
- [ ] Add some test cases that check the changes thoroughly including negative and positive cases if possible
- [ ] Add screenshots for manual tests if appropriate
- [ ] [Run test](https://kyuubi.readthedocs.io/en/master/develop_tools/testing.html#running-tests) locally before make a pull request
Closes #4369 from pan3793/spark-3.3.2.
Closes #4338
3232bf9f [Cheng Pan] [KYUUBI #4338][FOLLOWUP] Fix K8s integration tests
Authored-by: Cheng Pan <ch...@apache.org>
Signed-off-by: Cheng Pan <ch...@apache.org>
---
.github/workflows/master.yml | 1 -
.../kubernetes/test/deployment/KyuubiOnKubernetesTestsSuite.scala | 2 +-
.../kyuubi/kubernetes/test/spark/SparkOnKubernetesTestsSuite.scala | 2 +-
3 files changed, 2 insertions(+), 3 deletions(-)
diff --git a/.github/workflows/master.yml b/.github/workflows/master.yml
index b0ab493a4..a70117826 100644
--- a/.github/workflows/master.yml
+++ b/.github/workflows/master.yml
@@ -434,7 +434,6 @@ jobs:
run: >-
./build/mvn ${MVN_OPT} clean install
-Pflink-provided,hive-provided
- -Pspark-3.2
-Pkubernetes-it
-Dtest=none -DwildcardSuites=org.apache.kyuubi.kubernetes.test.spark
- name: Print Driver Pod logs
diff --git a/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/deployment/KyuubiOnKubernetesTestsSuite.scala b/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/deployment/KyuubiOnKubernetesTestsSuite.scala
index 6cf8dfcc8..bc7c98a80 100644
--- a/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/deployment/KyuubiOnKubernetesTestsSuite.scala
+++ b/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/deployment/KyuubiOnKubernetesTestsSuite.scala
@@ -54,7 +54,7 @@ class KyuubiOnKubernetesWithSparkTestsBase extends WithKyuubiServerOnKubernetes
super.connectionConf ++
Map(
"spark.master" -> s"k8s://$miniKubeApiMaster",
- "spark.kubernetes.container.image" -> "apache/spark:3.3.2",
+ "spark.kubernetes.container.image" -> "apache/spark:v3.3.2",
"spark.executor.memory" -> "512M",
"spark.driver.memory" -> "1024M",
"spark.kubernetes.driver.request.cores" -> "250m",
diff --git a/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/spark/SparkOnKubernetesTestsSuite.scala b/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/spark/SparkOnKubernetesTestsSuite.scala
index 798618e4c..019de840d 100644
--- a/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/spark/SparkOnKubernetesTestsSuite.scala
+++ b/integration-tests/kyuubi-kubernetes-it/src/test/scala/org/apache/kyuubi/kubernetes/test/spark/SparkOnKubernetesTestsSuite.scala
@@ -45,7 +45,7 @@ abstract class SparkOnKubernetesSuiteBase
// TODO Support more Spark version
// Spark official docker image: https://hub.docker.com/r/apache/spark/tags
KyuubiConf().set("spark.master", s"k8s://$apiServerAddress")
- .set("spark.kubernetes.container.image", "apache/spark:v3.2.1")
+ .set("spark.kubernetes.container.image", "apache/spark:v3.3.2")
.set("spark.kubernetes.container.image.pullPolicy", "IfNotPresent")
.set("spark.executor.instances", "1")
.set("spark.executor.memory", "512M")