You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by gu...@apache.org on 2021/11/08 00:30:49 UTC

[spark] branch master updated: [SPARK-37120][BUILD] Add Daily GitHub Action jobs for Java11/17

This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new 442dedb  [SPARK-37120][BUILD] Add Daily GitHub Action jobs for Java11/17
442dedb is described below

commit 442dedba835f43532d049adb8b56ba05bf675f3d
Author: Dongjoon Hyun <do...@apache.org>
AuthorDate: Mon Nov 8 09:30:07 2021 +0900

    [SPARK-37120][BUILD] Add Daily GitHub Action jobs for Java11/17
    
    ### What changes were proposed in this pull request?
    
    This PR aims to add Daily GitHub Action jobs for Java 11/17.
    
    ### Why are the changes needed?
    
    To add a test coverage on Java 11/17.
    
    ### Does this PR introduce _any_ user-facing change?
    
    No.
    
    ### How was this patch tested?
    
    N/A.
    
    Closes #34508 from dongjoon-hyun/SPARK-37120.
    
    Authored-by: Dongjoon Hyun <do...@apache.org>
    Signed-off-by: Hyukjin Kwon <gu...@apache.org>
---
 .github/workflows/build_and_test.yml | 24 +++++++++++++++++++++++-
 1 file changed, 23 insertions(+), 1 deletion(-)

diff --git a/.github/workflows/build_and_test.yml b/.github/workflows/build_and_test.yml
index 3f2d500..9f375b3 100644
--- a/.github/workflows/build_and_test.yml
+++ b/.github/workflows/build_and_test.yml
@@ -33,12 +33,17 @@ on:
     - cron: '0 7 * * *'
     # PySpark coverage for master branch
     - cron: '0 10 * * *'
+    # Java 11
+    - cron: '0 13 * * *'
+    # Java 17
+    - cron: '0 16 * * *'
 
 jobs:
   configure-jobs:
     name: Configure jobs
     runs-on: ubuntu-20.04
     outputs:
+      java: ${{ steps.set-outputs.outputs.java }}
       branch: ${{ steps.set-outputs.outputs.branch }}
       hadoop: ${{ steps.set-outputs.outputs.hadoop }}
       type: ${{ steps.set-outputs.outputs.type }}
@@ -48,26 +53,43 @@ jobs:
       id: set-outputs
       run: |
         if [ "${{ github.event.schedule }}" = "0 1 * * *" ]; then
+          echo '::set-output name=java::8'
           echo '::set-output name=branch::master'
           echo '::set-output name=type::scheduled'
           echo '::set-output name=envs::{}'
           echo '::set-output name=hadoop::hadoop2.7'
         elif [ "${{ github.event.schedule }}" = "0 4 * * *" ]; then
+          echo '::set-output name=java::8'
           echo '::set-output name=branch::master'
           echo '::set-output name=type::scheduled'
           echo '::set-output name=envs::{"SCALA_PROFILE": "scala2.13"}'
           echo '::set-output name=hadoop::hadoop3.2'
         elif [ "${{ github.event.schedule }}" = "0 7 * * *" ]; then
+          echo '::set-output name=java::8'
           echo '::set-output name=branch::branch-3.2'
           echo '::set-output name=type::scheduled'
           echo '::set-output name=envs::{"SCALA_PROFILE": "scala2.13"}'
           echo '::set-output name=hadoop::hadoop3.2'
         elif [ "${{ github.event.schedule }}" = "0 10 * * *" ]; then
+          echo '::set-output name=java::8'
           echo '::set-output name=branch::master'
           echo '::set-output name=type::pyspark-coverage-scheduled'
           echo '::set-output name=envs::{"PYSPARK_CODECOV": "true"}'
           echo '::set-output name=hadoop::hadoop3.2'
+        elif [ "${{ github.event.schedule }}" = "0 13 * * *" ]; then
+          echo '::set-output name=java::11'
+          echo '::set-output name=branch::branch-3.2'
+          echo '::set-output name=type::scheduled'
+          echo '::set-output name=envs::{}'
+          echo '::set-output name=hadoop::hadoop3.2'
+        elif [ "${{ github.event.schedule }}" = "0 16 * * *" ]; then
+          echo '::set-output name=java::17'
+          echo '::set-output name=branch::branch-3.2'
+          echo '::set-output name=type::scheduled'
+          echo '::set-output name=envs::{}'
+          echo '::set-output name=hadoop::hadoop3.2'
         else
+          echo '::set-output name=java::8'
           echo '::set-output name=branch::master' # Default branch to run on. CHANGE here when a branch is cut out.
           echo '::set-output name=type::regular'
           echo '::set-output name=envs::{}'
@@ -89,7 +111,7 @@ jobs:
       fail-fast: false
       matrix:
         java:
-          - 8
+          - ${{ needs.configure-jobs.outputs.java }}
         hadoop:
           - ${{ needs.configure-jobs.outputs.hadoop }}
         hive:

---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org