You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@flink.apache.org by ch...@apache.org on 2022/03/14 11:25:02 UTC
[flink] 04/04: disable azure
This is an automated email from the ASF dual-hosted git repository.
chesnay pushed a commit to branch experiment_gha_docs
in repository https://gitbox.apache.org/repos/asf/flink.git
commit 89cd9dcb583ae80f3ba89e972dfabbe061004592
Author: Chesnay Schepler <ch...@apache.org>
AuthorDate: Mon Mar 14 12:21:59 2022 +0100
disable azure
---
azure-pipelines.yml | 98 -----------------------------------------------------
1 file changed, 98 deletions(-)
diff --git a/azure-pipelines.yml b/azure-pipelines.yml
deleted file mode 100644
index 649643f..0000000
--- a/azure-pipelines.yml
+++ /dev/null
@@ -1,98 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements. See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-#
-# This file defines an Azure Pipeline build for testing Flink. It is intended to be used
-# with a free Azure Pipelines account.
-# It has the following features:
-# - default builds for pushes / pull requests
-# - end-to-end tests
-#
-#
-# For the "apache/flink" repository, we are using the pipeline definition located in
-# tools/azure-pipelines/build-apache-repo.yml
-# That file points to custom, self-hosted build agents for faster pull request build processing and
-# integration with Flinkbot.
-# The custom pipeline definition file is configured in the "Pipeline settings" screen
-# of the Azure Pipelines web ui.
-#
-
-trigger:
- branches:
- include:
- - '*' # must quote since "*" is a YAML reserved character; we want a string
-
-resources:
- containers:
- # Container with Maven 3.2.5, SSL to have the same environment everywhere.
- # see https://github.com/flink-ci/flink-ci-docker
- - container: flink-build-container
- image: chesnay/flink-ci:java_8_11_17
- # On AZP provided machines, set this flag to allow writing coredumps in docker
- options: --privileged
-
-# Define variables:
-# - See tools/azure-pipelines/jobs-template.yml for a short summary of the caching
-# - See https://stackoverflow.com/questions/60742105/how-can-i-access-a-secret-value-from-an-azure-pipelines-expression
-# to understand why the secrets are handled like this
-variables:
- MAVEN_CACHE_FOLDER: $(Pipeline.Workspace)/.m2/repository
- E2E_CACHE_FOLDER: $(Pipeline.Workspace)/e2e_cache
- E2E_TARBALL_CACHE: $(Pipeline.Workspace)/e2e_artifact_cache
- MAVEN_ARGS: '-Dmaven.repo.local=$(MAVEN_CACHE_FOLDER)'
- CACHE_KEY: maven | $(Agent.OS) | **/pom.xml, !**/target/**
- CACHE_FALLBACK_KEY: maven | $(Agent.OS)
- DOCKER_IMAGES_CACHE_KEY: docker-images-cache | $(Agent.OS) | **/cache_docker_images.sh | flink-test-utils-parent/**/DockerImageVersions.java
- DOCKER_IMAGES_CACHE_FOLDER: $(Pipeline.Workspace)/.docker-cache
- FLINK_ARTIFACT_DIR: $(Pipeline.Workspace)/flink_artifact
- SECRET_S3_BUCKET: $[variables.IT_CASE_S3_BUCKET]
- SECRET_S3_ACCESS_KEY: $[variables.IT_CASE_S3_ACCESS_KEY]
- SECRET_S3_SECRET_KEY: $[variables.IT_CASE_S3_SECRET_KEY]
- SECRET_GLUE_SCHEMA_ACCESS_KEY: $[variables.IT_CASE_GLUE_SCHEMA_ACCESS_KEY]
- SECRET_GLUE_SCHEMA_SECRET_KEY: $[variables.IT_CASE_GLUE_SCHEMA_SECRET_KEY]
-
-
-stages:
- # CI / PR triggered stage:
- - stage: ci
- displayName: "CI build (custom builders)"
- condition: not(eq(variables['MODE'], 'release'))
- jobs:
- - template: tools/azure-pipelines/jobs-template.yml
- parameters: # see template file for a definition of the parameters.
- stage_name: ci_build
- test_pool_definition:
- vmImage: 'ubuntu-20.04'
- e2e_pool_definition:
- vmImage: 'ubuntu-20.04'
- environment: PROFILE="-Dhadoop.version=2.8.5 -Dinclude_hadoop_aws -Dscala-2.12"
- run_end_to_end: false
- container: flink-build-container
- jdk: 8
- - job: docs_404_check # run on a MSFT provided machine
- pool:
- vmImage: 'ubuntu-20.04'
- steps:
- - script: ./tools/ci/docs.sh
- # CI / Special stage for release, e.g. building PyFlink wheel packages, etc:
- - stage: ci_release
- displayName: "CI build (release)"
- condition: and(eq(variables['Build.Reason'], 'Manual'), eq(variables['MODE'], 'release'))
- jobs:
- - template: tools/azure-pipelines/build-python-wheels.yml
- parameters:
- stage_name: cron_python_wheels
- environment: PROFILE="-Dhadoop.version=2.8.5 -Dinclude_hadoop_aws -Dscala-2.12"
- container: flink-build-container