You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/07/25 19:08:44 UTC

Build failed in Jenkins: beam_PostCommit_XVR_Dataflow #992

See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/992/display/redirect>

Changes:


------------------------------------------
[...truncated 1.77 MB...]
soupsieve==2.2.1
tenacity==8.0.1
tensorboard==2.5.0
tensorboard-data-server==0.6.1
tensorboard-plugin-wit==1.8.0
tensorflow==2.5.0
tensorflow-estimator==2.5.0
termcolor==1.1.0
threadpoolctl==2.2.0
typing-extensions==3.7.4.3
uritemplate==3.0.1
urllib3==1.25.11
wcwidth==0.2.5
Werkzeug==2.0.1
wheel==0.36.2
wrapt==1.12.1
Removing intermediate container 4ce950051131
 ---> 8869d37fe45e
Step 22/27 : RUN pip check
 ---> Running in 50f6097cc47b
No broken requirements found.
Removing intermediate container 50f6097cc47b
 ---> 283cc3538749
Step 23/27 : COPY target/LICENSE /opt/apache/beam/
 ---> fee93acaf50c
Step 24/27 : COPY target/LICENSE.python /opt/apache/beam/
 ---> a0fd433d070b
Step 25/27 : COPY target/NOTICE /opt/apache/beam/
 ---> 92672243319c
Step 26/27 : ADD target/launcher/linux_amd64/boot /opt/apache/beam/
 ---> d71d59aff19e
Step 27/27 : ENTRYPOINT ["/opt/apache/beam/boot"]
 ---> Running in 82e7e8fc1856
Removing intermediate container 82e7e8fc1856
 ---> 5755bf27e7f6
Successfully built 5755bf27e7f6
Successfully tagged apache/beam_python3.8_sdk:2.33.0.dev

> Task :runners:google-cloud-dataflow-java:validatesCrossLanguageRunnerSetup
Launching Java expansion service @ 36569
Launching Python expansion service @ 33397

> Task :runners:google-cloud-dataflow-java:validatesCrossLanguageRunnerPythonUsingSql
>>> RUNNING integration tests with pipeline options: --runner=TestDataflowRunner --project=apache-beam-testing --region=us-central1 --sdk_harness_container_image_overrides=.*java.*,us.gcr.io/apache-beam-testing/java-postcommit-it/java:20210725184910 --experiments=use_runner_v2 --experiments=shuffle_mode=appliance --sdk_location=<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/ws/src/sdks/python/build/apache-beam.tar.gz>
>>>   pytest options: --capture=no --numprocesses=8 --timeout=4500 --log-cli-level=INFO
>>>   collect markers: -m=xlang_sql_expansion_service
============================= test session starts ==============================
platform linux -- Python 3.6.8, pytest-4.6.11, py-1.10.0, pluggy-0.13.1
rootdir: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/ws/src/sdks/python,> inifile: pytest.ini
plugins: xdist-1.34.0, timeout-1.4.2, forked-1.3.0, requests-mock-1.9.3
timeout: 4500.0s
timeout method: signal
timeout func_only: False
gw0 I / gw1 I / gw2 I / gw3 I / gw4 I / gw5 I / gw6 I / gw7 I
[gw0] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
[gw1] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
[gw3] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
[gw2] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
[gw4] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
[gw5] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
[gw6] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
[gw7] Python 3.6.8 (default, Dec 24 2018, 19:24:27)  -- [GCC 5.4.0 20160609]
gw0 [9] / gw1 [9] / gw2 [9] / gw3 [9] / gw4 [9] / gw5 [9] / gw6 [9] / gw7 [9]

scheduling tests via LoadScheduling

apache_beam/transforms/sql_test.py::SqlTransformTest::test_generate_data 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_windowing_before_sql 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_tagged_join 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_row 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_project 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_filter 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_map 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_agg 
[gw4] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_project 
[gw5] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_row 
[gw1] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_filter 
[gw7] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_windowing_before_sql 
[gw0] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_agg 
apache_beam/transforms/sql_test.py::SqlTransformTest::test_zetasql_generate_data 
[gw3] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_generate_data 
[gw6] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_tagged_join 
[gw2] PASSED apache_beam/transforms/sql_test.py::SqlTransformTest::test_map 
[gw0] FAILED apache_beam/transforms/sql_test.py::SqlTransformTest::test_zetasql_generate_data 

=================================== FAILURES ===================================
_________________ SqlTransformTest.test_zetasql_generate_data __________________
[gw0] linux -- Python 3.6.8 <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/ws/src/build/gradleenv/1922375555/bin/python3.6>

self = <apache_beam.transforms.sql_test.SqlTransformTest testMethod=test_zetasql_generate_data>

    def test_zetasql_generate_data(self):
      with TestPipeline() as p:
        out = p | SqlTransform(
            """SELECT
              CAST(1 AS INT64) AS `int`,
              CAST('foo' AS STRING) AS `str`,
              CAST(3.14  AS FLOAT64) AS `flt`""",
            dialect="zetasql")
>       assert_that(out, equal_to([(1, "foo", 3.14)]))

apache_beam/transforms/sql_test.py:160: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:585: in __exit__
    self.result = self.run()
apache_beam/testing/test_pipeline.py:114: in run
    False if self.not_use_test_runner_api else test_runner_api))
apache_beam/pipeline.py:564: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:54: in run_pipeline
    self).run_pipeline(pipeline, options)
apache_beam/runners/dataflow/dataflow_runner.py:588: in run_pipeline
    self.dataflow_client.create_job(self.job), self)
apache_beam/utils/retry.py:253: in wrapper
    return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:688: in create_job
    return self.submit_job_description(job)
apache_beam/utils/retry.py:253: in wrapper
    return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:789: in submit_job_description
    response = self._client.projects_locations_jobs.Create(request)
apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py:903: in Create
    config, request, global_params=global_params)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/apitools/base/py/base_api.py:731: in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
../../build/gradleenv/1922375555/lib/python3.6/site-packages/apitools/base/py/base_api.py:737: in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <apache_beam.runners.dataflow.internal.clients.dataflow.dataflow_v1b3_client.DataflowV1b3.ProjectsLocationsJobsService object at 0x7f0169bfc0f0>
method_config = <ApiMethodInfo
 relative_path: 'v1b3/projects/{projectId}/locations/{location}/jobs'
 method_id: 'dataflow.projects.lo...DataflowProjectsLocationsJobsCreateRequest'
 response_type_name: 'Job'
 request_field: 'job'
 supports_download: False>
http_response = Response(info={'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Sun, 2...', request_url='https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json')
request = <DataflowProjectsLocationsJobsCreateRequest
 job: <Job
 environment: <Environment
 clusterManagerApiService: 'compute....empFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
 location: 'us-central1'
 projectId: 'apache-beam-testing'>

    def __ProcessHttpResponse(self, method_config, http_response, request):
        """Process the given http response."""
        if http_response.status_code not in (http_client.OK,
                                             http_client.CREATED,
                                             http_client.NO_CONTENT):
            raise exceptions.HttpError.FromResponse(
>               http_response, method_config=method_config, request=request)
E           apitools.base.py.exceptions.HttpConflictError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Sun, 25 Jul 2021 19:08:38 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '409', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{
E             "error": {
E               "code": 409,
E               "message": "(cd8616b8b1d7c856): The workflow could not be created. Causes: (edbe3f79ba05332e): There is already an active job named beamapp-jenkins-0725190759-188570. If you want to  submit a second job, try again by setting a different name.",
E               "status": "ALREADY_EXISTS"
E             }
E           }
E           >

../../build/gradleenv/1922375555/lib/python3.6/site-packages/apitools/base/py/base_api.py:604: HttpConflictError
------------------------------ Captured log call -------------------------------
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:231 Using pre-built snapshot at <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.33.0-SNAPSHOT.jar>
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:113 Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/ws/src/sdks/java/extensions/sql/expansion-service/build/libs/beam-sdks-java-extensions-sql-expansion-service-2.33.0-SNAPSHOT.jar'> '36055']
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'Starting expansion service at localhost:36055'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'Jul 25, 2021 7:07:48 PM org.apache.beam.sdk.expansion.service.ExpansionService loadRegisteredTransforms'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'INFO: Registering external transforms: [beam:external:java:sql:v1, beam:external:java:generate_sequence:v1]'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'\tbeam:external:java:sql:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$47/93314457@6591f517'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'\tbeam:external:java:generate_sequence:v1: org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader$$Lambda$47/93314457@345965f2'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'Jul 25, 2021 7:07:50 PM org.apache.beam.sdk.expansion.service.ExpansionService expand'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b"INFO: Expanding 'SqlTransform(beam:external:java:sql:v1)' with URN 'beam:external:java:sql:v1'"
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'Jul 25, 2021 7:07:51 PM org.apache.beam.sdk.expansion.service.ExpansionService$ExternalTransformRegistrarLoader payloadToConfig'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b"WARNING: Configuration class 'org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar$Configuration' has no schema registered. Attempting to construct with setter approach."
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'Jul 25, 2021 7:07:54 PM org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner convertToBeamRelInternal'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'INFO: BEAMPlan>'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b"BeamZetaSqlCalcRel(expr#0=[{inputs}], expr#1=[1:BIGINT], expr#2=['foo':VARCHAR], expr#3=[3.1400000000000001243E0:DOUBLE], int=[$t1], str=[$t2], flt=[$t3])"
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b'  BeamValuesRel(tuples=[[{ 0 }]])'
INFO     apache_beam.utils.subprocess_server:subprocess_server.py:122 b''
INFO     apache_beam.runners.portability.stager:stager.py:300 Copying Beam SDK "<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/ws/src/sdks/python/build/apache-beam.tar.gz"> to staging location.
WARNING  root:environments.py:374 Make sure that locally built Python SDK docker image has Python 3.6 interpreter.
INFO     root:environments.py:380 Default Python SDK image for environment is apache/beam_python3.6_sdk:2.33.0.dev
INFO     root:environments.py:296 Using provided Python SDK container image: gcr.io/cloud-dataflow/v1beta3/python36-fnapi:beam-master-20210720
INFO     root:environments.py:304 Python SDK container image set to "gcr.io/cloud-dataflow/v1beta3/python36-fnapi:beam-master-20210720" for Docker environment
INFO     apache_beam.runners.portability.fn_api_runner.translations:translations.py:650 ==================== <function pack_combiners at 0x7f016d7a5e18> ====================
INFO     apache_beam.runners.portability.fn_api_runner.translations:translations.py:650 ==================== <function sort_stages at 0x7f016d7ae620> ====================
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:452 Defaulting to the temp_location as staging_location: gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/icedtea-sound-Bdoi2wYa757-fzq5vconCy4SSQ22ZaOq7yuC98fKPs8.jar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/icedtea-sound-Bdoi2wYa757-fzq5vconCy4SSQ22ZaOq7yuC98fKPs8.jar in 0 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/jaccess-CMbK-IOdQPLKHEqCuDnE-yBk-VpbtVT-hgjbHRUGO78.jar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/jaccess-CMbK-IOdQPLKHEqCuDnE-yBk-VpbtVT-hgjbHRUGO78.jar in 0 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/localedata-ae5Z0L6ak4922fztWeWy7ajiWXdG3ubNrwerJRFoFj0.jar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/localedata-ae5Z0L6ak4922fztWeWy7ajiWXdG3ubNrwerJRFoFj0.jar in 0 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/nashorn-XHtz_UehGpYcLTOrATrTnMNVUgEVa_ttoWkPxnVfqTo.jar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/nashorn-XHtz_UehGpYcLTOrATrTnMNVUgEVa_ttoWkPxnVfqTo.jar in 0 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/cldrdata-k07I6K9W3X5KTQbcDIEsqM0LXyM18f0eR6IaJw-P_xk.jar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/cldrdata-k07I6K9W3X5KTQbcDIEsqM0LXyM18f0eR6IaJw-P_xk.jar in 0 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/dnsns-RGhCDg3GVOQVC2r6ka2N0hmI4eqQH6VobuoAnQ74MnE.jar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/dnsns-RGhCDg3GVOQVC2r6ka2N0hmI4eqQH6VobuoAnQ74MnE.jar in 0 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/beam-sdks-java-extensions-sql-expansion-service-2.33.0-SNAPSHOT-4E0BC4cBgP4KLXuXHjtZeqrPEdvaCvqwC3WkgJqp_xg.jar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/beam-sdks-java-extensions-sql-expansion-service-2.33.0-SNAPSHOT-4E0BC4cBgP4KLXuXHjtZeqrPEdvaCvqwC3WkgJqp_xg.jar in 5 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/dataflow_python_sdk.tar...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/dataflow_python_sdk.tar in 0 seconds.
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:632 Starting GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/pipeline.pb...
INFO     apache_beam.runners.dataflow.internal.apiclient:apiclient.py:651 Completed GCS upload to gs://dataflow-staging-us-central1-77b801c0838aee13391c0d1885860494/beamapp-jenkins-0725190759-188570.1627240079.189109/pipeline.pb in 0 seconds.
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/ws/src/sdks/python/pytest_xlangSqlValidateRunner.xml> -
===================== 1 failed, 8 passed in 707.95 seconds =====================

> Task :runners:google-cloud-dataflow-java:validatesCrossLanguageRunnerPythonUsingSql FAILED

> Task :runners:google-cloud-dataflow-java:validatesCrossLanguageRunnerCleanup
Stopping expansion service pid: 14313.
Stopping expansion service pid: 14316.

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20210725184910
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4482a0514e567d4217117ceaaab14e1ef29b43c34baf3cf383e654d165d0c95a
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4482a0514e567d4217117ceaaab14e1ef29b43c34baf3cf383e654d165d0c95a
  Associated tags:
 - 20210725184910
Tags:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java:20210725184910
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20210725184910].
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4482a0514e567d4217117ceaaab14e1ef29b43c34baf3cf383e654d165d0c95a].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 19m 35s
143 actionable tasks: 51 executed, 13 from cache, 79 up-to-date

Publishing build scan...
https://gradle.com/s/7eylb3bnbio26

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_XVR_Dataflow #993

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_XVR_Dataflow/993/display/redirect>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org