You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@flink.apache.org by ch...@apache.org on 2022/03/14 11:25:02 UTC

[flink] 04/04: disable azure

This is an automated email from the ASF dual-hosted git repository.

chesnay pushed a commit to branch experiment_gha_docs
in repository https://gitbox.apache.org/repos/asf/flink.git

commit 89cd9dcb583ae80f3ba89e972dfabbe061004592
Author: Chesnay Schepler <ch...@apache.org>
AuthorDate: Mon Mar 14 12:21:59 2022 +0100

    disable azure
---
 azure-pipelines.yml | 98 -----------------------------------------------------
 1 file changed, 98 deletions(-)

diff --git a/azure-pipelines.yml b/azure-pipelines.yml
deleted file mode 100644
index 649643f..0000000
--- a/azure-pipelines.yml
+++ /dev/null
@@ -1,98 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#    http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-#
-# This file defines an Azure Pipeline build for testing Flink. It is intended to be used
-# with a free Azure Pipelines account.
-# It has the following features:
-#  - default builds for pushes / pull requests
-#  - end-to-end tests
-#
-#
-# For the "apache/flink" repository, we are using the pipeline definition located in
-#   tools/azure-pipelines/build-apache-repo.yml
-# That file points to custom, self-hosted build agents for faster pull request build processing and 
-# integration with Flinkbot.
-# The custom pipeline definition file is configured in the "Pipeline settings" screen
-# of the Azure Pipelines web ui.
-#
-
-trigger:
-  branches:
-    include:
-    - '*'  # must quote since "*" is a YAML reserved character; we want a string
-
-resources:
-  containers:
-  # Container with Maven 3.2.5, SSL to have the same environment everywhere.
-  # see https://github.com/flink-ci/flink-ci-docker
-  - container: flink-build-container
-    image: chesnay/flink-ci:java_8_11_17
-    # On AZP provided machines, set this flag to allow writing coredumps in docker
-    options: --privileged
-
-# Define variables:
-# - See tools/azure-pipelines/jobs-template.yml for a short summary of the caching
-# - See https://stackoverflow.com/questions/60742105/how-can-i-access-a-secret-value-from-an-azure-pipelines-expression
-#   to understand why the secrets are handled like this
-variables:
-  MAVEN_CACHE_FOLDER: $(Pipeline.Workspace)/.m2/repository
-  E2E_CACHE_FOLDER: $(Pipeline.Workspace)/e2e_cache
-  E2E_TARBALL_CACHE: $(Pipeline.Workspace)/e2e_artifact_cache
-  MAVEN_ARGS: '-Dmaven.repo.local=$(MAVEN_CACHE_FOLDER)'
-  CACHE_KEY: maven | $(Agent.OS) | **/pom.xml, !**/target/**
-  CACHE_FALLBACK_KEY: maven | $(Agent.OS)
-  DOCKER_IMAGES_CACHE_KEY: docker-images-cache | $(Agent.OS) | **/cache_docker_images.sh | flink-test-utils-parent/**/DockerImageVersions.java
-  DOCKER_IMAGES_CACHE_FOLDER: $(Pipeline.Workspace)/.docker-cache
-  FLINK_ARTIFACT_DIR: $(Pipeline.Workspace)/flink_artifact
-  SECRET_S3_BUCKET: $[variables.IT_CASE_S3_BUCKET]
-  SECRET_S3_ACCESS_KEY: $[variables.IT_CASE_S3_ACCESS_KEY]
-  SECRET_S3_SECRET_KEY: $[variables.IT_CASE_S3_SECRET_KEY]
-  SECRET_GLUE_SCHEMA_ACCESS_KEY: $[variables.IT_CASE_GLUE_SCHEMA_ACCESS_KEY]
-  SECRET_GLUE_SCHEMA_SECRET_KEY: $[variables.IT_CASE_GLUE_SCHEMA_SECRET_KEY]
-
-
-stages:
-  # CI / PR triggered stage:
-  - stage: ci
-    displayName: "CI build (custom builders)"
-    condition: not(eq(variables['MODE'], 'release'))
-    jobs:
-      - template: tools/azure-pipelines/jobs-template.yml
-        parameters: # see template file for a definition of the parameters.
-          stage_name: ci_build
-          test_pool_definition:
-            vmImage: 'ubuntu-20.04'
-          e2e_pool_definition:
-            vmImage: 'ubuntu-20.04'
-          environment: PROFILE="-Dhadoop.version=2.8.5 -Dinclude_hadoop_aws -Dscala-2.12"
-          run_end_to_end: false
-          container: flink-build-container
-          jdk: 8
-      - job: docs_404_check # run on a MSFT provided machine
-        pool:
-          vmImage: 'ubuntu-20.04'
-        steps:
-          - script: ./tools/ci/docs.sh
-  # CI / Special stage for release, e.g. building PyFlink wheel packages, etc:
-  - stage: ci_release
-    displayName: "CI build (release)"
-    condition: and(eq(variables['Build.Reason'], 'Manual'), eq(variables['MODE'], 'release'))
-    jobs:
-      - template: tools/azure-pipelines/build-python-wheels.yml
-        parameters:
-          stage_name: cron_python_wheels
-          environment: PROFILE="-Dhadoop.version=2.8.5 -Dinclude_hadoop_aws -Dscala-2.12"
-          container: flink-build-container