You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2023/10/15 11:18:14 UTC

Build failed in Jenkins: beam_PostCommit_Python38 #4711

See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4711/display/redirect>

Changes:


------------------------------------------
[...truncated 12.38 MB...]
popenargs = (['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpcmdv870n', ...],)
kwargs = {'stdout': -1}, process = <subprocess.Popen object at 0x7fcfe682e430>
stdout = b'', stderr = None, retcode = 1

    def run(*popenargs,
            input=None, capture_output=False, timeout=None, check=False, **kwargs):
        """Run command with arguments and return a CompletedProcess instance.
    
        The returned instance will have attributes args, returncode, stdout and
        stderr. By default, stdout and stderr are not captured, and those attributes
        will be None. Pass stdout=PIPE and/or stderr=PIPE in order to capture them.
    
        If check is True and the exit code was non-zero, it raises a
        CalledProcessError. The CalledProcessError object will have the return code
        in the returncode attribute, and output & stderr attributes if those streams
        were captured.
    
        If timeout is given, and the process takes too long, a TimeoutExpired
        exception will be raised.
    
        There is an optional argument "input", allowing you to
        pass bytes or a string to the subprocess's stdin.  If you use this argument
        you may not also use the Popen constructor's "stdin" argument, as
        it will be used internally.
    
        By default, all communication is in bytes, and therefore any "input" should
        be bytes, and the stdout and stderr will be bytes. If in text mode, any
        "input" should be a string, and stdout and stderr will be strings decoded
        according to locale encoding, or by "encoding" if set. Text mode is
        triggered by setting any of text, encoding, errors or universal_newlines.
    
        The other arguments are the same as for the Popen constructor.
        """
        if input is not None:
            if kwargs.get('stdin') is not None:
                raise ValueError('stdin and input arguments may not both be used.')
            kwargs['stdin'] = PIPE
    
        if capture_output:
            if kwargs.get('stdout') is not None or kwargs.get('stderr') is not None:
                raise ValueError('stdout and stderr arguments may not be used '
                                 'with capture_output.')
            kwargs['stdout'] = PIPE
            kwargs['stderr'] = PIPE
    
        with Popen(*popenargs, **kwargs) as process:
            try:
                stdout, stderr = process.communicate(input, timeout=timeout)
            except TimeoutExpired as exc:
                process.kill()
                if _mswindows:
                    # Windows accumulates the output in a single blocking
                    # read() call run on child threads, with the timeout
                    # being done in a join() on those threads.  communicate()
                    # _after_ kill() is required to collect that and add it
                    # to the exception.
                    exc.stdout, exc.stderr = process.communicate()
                else:
                    # POSIX _communicate already populated the output so
                    # far into the TimeoutExpired exception.
                    process.wait()
                raise
            except:  # Including KeyboardInterrupt, communicate handled that.
                process.kill()
                # We don't call process.wait() as .__exit__ does that for us.
                raise
            retcode = process.poll()
            if check and retcode:
>               raise CalledProcessError(retcode, process.args,
                                         output=stdout, stderr=stderr)
E               subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpcmdv870n', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.

/usr/lib/python3.8/subprocess.py:516: CalledProcessError

During handling of the above exception, another exception occurred:

self = <apache_beam.examples.complete.juliaset.juliaset.juliaset_test_it.JuliaSetTestIT testMethod=test_run_example_with_setup_file>

    def test_run_example_with_setup_file(self):
      pipeline = TestPipeline(is_integration_test=True)
      coordinate_output = FileSystems.join(
          pipeline.get_option('output'),
          'juliaset-{}'.format(str(uuid.uuid4())),
          'coordinates.txt')
      extra_args = {
          'coordinate_output': coordinate_output,
          'grid_size': self.GRID_SIZE,
          'setup_file': os.path.normpath(
              os.path.join(os.path.dirname(__file__), '..', 'setup.py')),
          'on_success_matcher': all_of(PipelineStateMatcher(PipelineState.DONE)),
      }
      args = pipeline.get_full_options_as_args(**extra_args)
    
>     juliaset.run(args)

apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py:56: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/examples/complete/juliaset/juliaset/juliaset.py:116: in run
    (
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:53: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/dataflow/dataflow_runner.py:393: in run_pipeline
    artifacts = environments.python_sdk_dependencies(options)
apache_beam/transforms/environments.py:846: in python_sdk_dependencies
    return stager.Stager.create_job_resources(
apache_beam/runners/portability/stager.py:281: in create_job_resources
    tarball_file = Stager._build_setup_package(
apache_beam/runners/portability/stager.py:796: in _build_setup_package
    processes.check_output(build_setup_args)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

args = (['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpcmdv870n', ...],)
kwargs = {}

    def check_output(*args, **kwargs):
      if force_shell:
        kwargs['shell'] = True
      try:
        out = subprocess.check_output(*args, **kwargs)
      except OSError:
        raise RuntimeError("Executable {} not found".format(args[0]))
      except subprocess.CalledProcessError as error:
        if isinstance(args, tuple) and (args[0][2] == "pip"):
          raise RuntimeError( \
            "Full traceback: {} \n Pip install failed for package: {} \
            \n Output from execution of subprocess: {}" \
            .format(traceback.format_exc(), args[0][6], error.output))
        else:
>         raise RuntimeError("Full trace: {}, \
             output of the failed child process {} "\
            .format(traceback.format_exc(), error.output))
E         RuntimeError: Full trace: Traceback (most recent call last):
E           File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/processes.py",> line 89, in check_output
E             out = subprocess.check_output(*args, **kwargs)
E           File "/usr/lib/python3.8/subprocess.py", line 415, in check_output
E             return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
E           File "/usr/lib/python3.8/subprocess.py", line 516, in run
E             raise CalledProcessError(retcode, process.args,
E         subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpcmdv870n', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.
E         ,            output of the failed child process b''

apache_beam/utils/processes.py:99: RuntimeError
------------------------------ Captured log call -------------------------------
INFO     apache_beam.runners.portability.stager:stager.py:762 Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', '/tmp/tmpth8nx7o0/tmp_requirements.txt', '--exists-action', 'i', '--no-deps', '--implementation', 'cp', '--abi', 'cp38', '--platform', 'manylinux2014_x86_64']
INFO     apache_beam.runners.portability.stager:stager.py:795 Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpcmdv870n', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']>
=============================== warnings summary ===============================
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo?  You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
    @pytest.mark.uses_vertex_ai

apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    dataset_ref = client.dataset(unique_dataset_name, project=project)

apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
    return airline_df[at_top_airports].mean()

apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    table_ref = client.dataset(dataset_id).table(table_id)

apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
=========================== short test summary info ============================
FAILED apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py::JuliaSetTestIT::test_run_example_with_setup_file - RuntimeError: Full trace: Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/processes.py",> line 89, in check_output
    out = subprocess.check_output(*args, **kwargs)
  File "/usr/lib/python3.8/subprocess.py", line 415, in check_output
    return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
  File "/usr/lib/python3.8/subprocess.py", line 516, in run
    raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpcmdv870n', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.
,            output of the failed child process b''
====== 1 failed, 88 passed, 49 skipped, 17 warnings in 9040.76s (2:30:40) ======

> Task :sdks:python:test-suites:dataflow:py38:postCommitIT FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 448

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:vertexAIInferenceTest'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 3h 2m 46s
221 actionable tasks: 157 executed, 60 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/gircqildmoj5s

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Python38 #4718

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4718/display/redirect>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python38 #4717

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4717/display/redirect>

Changes:


------------------------------------------
[...truncated 12.04 MB...]
[gw1] PASSED apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it 
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_legacy_sql 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_multiple_destinations_transform 
apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform 
[gw5] PASSED apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_1M_python 
apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python 
[gw6] PASSED apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_multiple_destinations_transform 
apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail 
[gw3] PASSED apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read 
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read 
[gw0] PASSED apache_beam/dataframe/io_it_test.py::ReadUsingReadGbqTests::test_ReadGbq_with_computation 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write 
[gw1] PASSED apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_legacy_sql 
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform 
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads 
[gw5] PASSED apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy 
[gw5] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch 
[gw5] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_kms 
[gw5] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_kms 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token 
[gw5] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms 
[gw5] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token 
[gw5] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token 
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification 
[gw6] PASSED apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail 
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only 
[gw3] PASSED apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read 
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write 
[gw3] PASSED apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write 
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read 
[gw0] PASSED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting 
[gw1] PASSED apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types 
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro 
[gw5] PASSED apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification 
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection 
[gw3] PASSED apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read 
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts 
[gw2] PASSED apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source 
[gw0] PASSED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error 
[gw1] PASSED apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro 
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql 
[gw6] PASSED apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only 
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes 
[gw5] PASSED apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection 
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax 
[gw5] SKIPPED apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax 
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming 
[gw5] SKIPPED apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming 
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads 
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts 
[gw0] PASSED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source 
[gw5] PASSED apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification 
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model 
[gw1] PASSED apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql 
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint 
[gw3] PASSED apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts 
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context 
[gw3] SKIPPED apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context 
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification 
[gw3] SKIPPED apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch 
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch 
[gw3] SKIPPED apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch 
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it 
[gw3] SKIPPED apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it 
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution 
[gw3] SKIPPED apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution 
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs 
[gw3] SKIPPED apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs 
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement 
[gw3] SKIPPED apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement 
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it 
[gw4] PASSED apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit 
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances 
[gw6] PASSED apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes 
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item 
[gw0] PASSED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect 
[gw0] SKIPPED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema 
[gw1] PASSED apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint 
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm 
[gw1] SKIPPED apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm 
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model 
[gw1] SKIPPED apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model 
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn 
[gw1] SKIPPED apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn 
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2 
[gw1] SKIPPED apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2 
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2 
[gw1] SKIPPED apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve 
[gw5] PASSED apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model 
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression 
[gw3] PASSED apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it 
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it 
[gw3] SKIPPED apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it 
[gw4] PASSED apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances 
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs 
[gw6] PASSED apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item 
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event 
[gw5] PASSED apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression 
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification 
[gw0] PASSED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts 
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load 
[gw4] PASSED apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs 
[gw5] PASSED apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification 
[gw6] PASSED apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event 
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load 
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation 
[gw7] SKIPPED apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation 
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification 
[gw7] SKIPPED apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification 
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model 
[gw7] SKIPPED apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model 
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification 
[gw7] SKIPPED apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source 
[gw6] PASSED apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner 
[gw2] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner 

=============================== warnings summary ===============================
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo?  You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
    @pytest.mark.uses_vertex_ai

apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    dataset_ref = client.dataset(unique_dataset_name, project=project)

apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
    return airline_df[at_top_airports].mean()

apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    table_ref = client.dataset(dataset_id).table(table_id)

apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
========== 89 passed, 49 skipped, 17 warnings in 10246.12s (2:50:46) ===========

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 420

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:tensorRTtests'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 3h 24m 15s
221 actionable tasks: 160 executed, 57 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/el3et3qtoxv7k

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python38 #4716

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4716/display/redirect?page=changes>

Changes:

[noreply] Bump github.com/aws/aws-sdk-go-v2/feature/s3/manager in /sdks (#29002)


------------------------------------------
[...truncated 12.21 MB...]
        were captured.
    
        If timeout is given, and the process takes too long, a TimeoutExpired
        exception will be raised.
    
        There is an optional argument "input", allowing you to
        pass bytes or a string to the subprocess's stdin.  If you use this argument
        you may not also use the Popen constructor's "stdin" argument, as
        it will be used internally.
    
        By default, all communication is in bytes, and therefore any "input" should
        be bytes, and the stdout and stderr will be bytes. If in text mode, any
        "input" should be a string, and stdout and stderr will be strings decoded
        according to locale encoding, or by "encoding" if set. Text mode is
        triggered by setting any of text, encoding, errors or universal_newlines.
    
        The other arguments are the same as for the Popen constructor.
        """
        if input is not None:
            if kwargs.get('stdin') is not None:
                raise ValueError('stdin and input arguments may not both be used.')
            kwargs['stdin'] = PIPE
    
        if capture_output:
            if kwargs.get('stdout') is not None or kwargs.get('stderr') is not None:
                raise ValueError('stdout and stderr arguments may not be used '
                                 'with capture_output.')
            kwargs['stdout'] = PIPE
            kwargs['stderr'] = PIPE
    
        with Popen(*popenargs, **kwargs) as process:
            try:
                stdout, stderr = process.communicate(input, timeout=timeout)
            except TimeoutExpired as exc:
                process.kill()
                if _mswindows:
                    # Windows accumulates the output in a single blocking
                    # read() call run on child threads, with the timeout
                    # being done in a join() on those threads.  communicate()
                    # _after_ kill() is required to collect that and add it
                    # to the exception.
                    exc.stdout, exc.stderr = process.communicate()
                else:
                    # POSIX _communicate already populated the output so
                    # far into the TimeoutExpired exception.
                    process.wait()
                raise
            except:  # Including KeyboardInterrupt, communicate handled that.
                process.kill()
                # We don't call process.wait() as .__exit__ does that for us.
                raise
            retcode = process.poll()
            if check and retcode:
>               raise CalledProcessError(retcode, process.args,
                                         output=stdout, stderr=stderr)
E               subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpjywubp9j', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.

/usr/lib/python3.8/subprocess.py:516: CalledProcessError

During handling of the above exception, another exception occurred:

self = <apache_beam.examples.complete.juliaset.juliaset.juliaset_test_it.JuliaSetTestIT testMethod=test_run_example_with_setup_file>

    def test_run_example_with_setup_file(self):
      pipeline = TestPipeline(is_integration_test=True)
      coordinate_output = FileSystems.join(
          pipeline.get_option('output'),
          'juliaset-{}'.format(str(uuid.uuid4())),
          'coordinates.txt')
      extra_args = {
          'coordinate_output': coordinate_output,
          'grid_size': self.GRID_SIZE,
          'setup_file': os.path.normpath(
              os.path.join(os.path.dirname(__file__), '..', 'setup.py')),
          'on_success_matcher': all_of(PipelineStateMatcher(PipelineState.DONE)),
      }
      args = pipeline.get_full_options_as_args(**extra_args)
    
>     juliaset.run(args)

apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py:56: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/examples/complete/juliaset/juliaset/juliaset.py:116: in run
    (
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:53: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/dataflow/dataflow_runner.py:393: in run_pipeline
    artifacts = environments.python_sdk_dependencies(options)
apache_beam/transforms/environments.py:846: in python_sdk_dependencies
    return stager.Stager.create_job_resources(
apache_beam/runners/portability/stager.py:281: in create_job_resources
    tarball_file = Stager._build_setup_package(
apache_beam/runners/portability/stager.py:796: in _build_setup_package
    processes.check_output(build_setup_args)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

args = (['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpjywubp9j', ...],)
kwargs = {}

    def check_output(*args, **kwargs):
      if force_shell:
        kwargs['shell'] = True
      try:
        out = subprocess.check_output(*args, **kwargs)
      except OSError:
        raise RuntimeError("Executable {} not found".format(args[0]))
      except subprocess.CalledProcessError as error:
        if isinstance(args, tuple) and (args[0][2] == "pip"):
          raise RuntimeError( \
            "Full traceback: {} \n Pip install failed for package: {} \
            \n Output from execution of subprocess: {}" \
            .format(traceback.format_exc(), args[0][6], error.output))
        else:
>         raise RuntimeError("Full trace: {}, \
             output of the failed child process {} "\
            .format(traceback.format_exc(), error.output))
E         RuntimeError: Full trace: Traceback (most recent call last):
E           File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/processes.py",> line 89, in check_output
E             out = subprocess.check_output(*args, **kwargs)
E           File "/usr/lib/python3.8/subprocess.py", line 415, in check_output
E             return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
E           File "/usr/lib/python3.8/subprocess.py", line 516, in run
E             raise CalledProcessError(retcode, process.args,
E         subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpjywubp9j', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.
E         ,            output of the failed child process b''

apache_beam/utils/processes.py:99: RuntimeError
------------------------------ Captured log call -------------------------------
INFO     apache_beam.runners.portability.stager:stager.py:762 Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', '/tmp/tmpjntkj16h/tmp_requirements.txt', '--exists-action', 'i', '--no-deps', '--implementation', 'cp', '--abi', 'cp38', '--platform', 'manylinux2014_x86_64']
INFO     apache_beam.runners.portability.stager:stager.py:795 Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpjywubp9j', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']>
=============================== warnings summary ===============================
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo?  You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
    @pytest.mark.uses_vertex_ai

apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    dataset_ref = client.dataset(unique_dataset_name, project=project)

apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    table_ref = client.dataset(dataset_id).table(table_id)

apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
    return airline_df[at_top_airports].mean()

apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
=========================== short test summary info ============================
FAILED apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py::JuliaSetTestIT::test_run_example_with_setup_file - RuntimeError: Full trace: Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/processes.py",> line 89, in check_output
    out = subprocess.check_output(*args, **kwargs)
  File "/usr/lib/python3.8/subprocess.py", line 415, in check_output
    return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
  File "/usr/lib/python3.8/subprocess.py", line 516, in run
    raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmpjywubp9j', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.
,            output of the failed child process b''
====== 1 failed, 88 passed, 49 skipped, 17 warnings in 8901.58s (2:28:21) ======

> Task :sdks:python:test-suites:dataflow:py38:postCommitIT FAILED

FAILURE: Build completed with 3 failures.

1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 420

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:tensorRTtests'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52

* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================

3: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 3h 2m
221 actionable tasks: 157 executed, 60 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/f7p4h4tpmnhpy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python38 #4715

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4715/display/redirect>

Changes:


------------------------------------------
[...truncated 9.75 MB...]
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1697445807
  nanos: 839660882
}
message: "Python sdk harness exiting."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker_main.py:211"
thread: "MainThread"

fe0190dbe899a4211748276d39366d2ef23e3b387b9e7c179c50f2a484bf0cfe
INFO:apache_beam.runners.portability.local_job_service:Completed job in 30.189043283462524 seconds with state DONE.
INFO:root:Completed job in 30.189043283462524 seconds with state DONE.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE

> Task :sdks:python:test-suites:portable:py38:portableWordCountSparkRunnerBatch
INFO:apache_beam.runners.worker.worker_pool_main:Listening for workers at localhost:35351
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7fd978f67e50> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7fd978f67ee0> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7fd978f69670> ====================
INFO:apache_beam.utils.subprocess_server:Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/runners/spark/3/job-server/build/libs/beam-runners-spark-3-job-server-2.52.0-SNAPSHOT.jar'> '--spark-master-url' 'local[4]' '--artifacts-dir' '/tmp/beam-temp4w3dsuc1/artifactsxh911w96' '--job-port' '46267' '--artifact-port' '0' '--expansion-port' '0']
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:42 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:41147
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:40587
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:46267
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
WARNING:root:Waiting for grpc channel to be ready at localhost:46267.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_c34b8b86-8e40-43f2-b419-13a2bd3eb75a.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_c34b8b86-8e40-43f2-b419-13a2bd3eb75a.ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_c34b8b86-8e40-43f2-b419-13a2bd3eb75a.null.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_c34b8b86-8e40-43f2-b419-13a2bd3eb75a.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:46 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-1016084346-b11c4e7f_f6747402-e829-4173-80c1-1c1b51e36794
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:46 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-1016084346-b11c4e7f_f6747402-e829-4173-80c1-1c1b51e36794
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
  with Pipeline() as p:
    p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:46 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:47 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.util.log: Logging initialized @8363ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_382-8u382-ga-1~20.04.1-b05
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.Server: Started @8485ms
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@2cff1953{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6c02adbd{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@56bcbe0{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@19fb336d{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@297b0fca{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3ef124c8{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3c501da0{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3bd5446a{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5aa8f8cb{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6b498289{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@45d868b1{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4c939b6e{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4f48f69d{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3c1033b1{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@61f53dd4{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@74885ed7{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5b2eaced{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@675b2350{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@337d7a0a{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@41d7b199{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@53d66583{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@12802866{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2aeae892{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6a87c617{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3da73572{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@d8e42f1{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@26b839a9{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:48 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-1016084346-b11c4e7f_f6747402-e829-4173-80c1-1c1b51e36794 on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:46099.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:35895.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:38225
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:51 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1016084346-b11c4e7f_f6747402-e829-4173-80c1-1c1b51e36794: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1016084346-b11c4e7f_f6747402-e829-4173-80c1-1c1b51e36794 finished.
INFO:apache_beam.utils.subprocess_server:23/10/16 08:43:52 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@2cff1953{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread read_state:
Traceback (most recent call last):
  File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
Exception in thread run_worker_1-1:
Traceback (most recent call last):
  File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
    self.run()
  File "/usr/lib/python3.8/threading.py", line 870, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1035, in pull_responses
    self.run()
  File "/usr/lib/python3.8/threading.py", line 870, in run
    for response in responses:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 262, in run
    for work_request in self._control_stub.Control(get_responses()):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:35895 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-16T08:43:53.069684452+00:00"}"
>
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38225 {created_time:"2023-10-16T08:43:53.069675919+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
  File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:46099 {created_time:"2023-10-16T08:43:53.069701391+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
    self.run()
  File "/usr/lib/python3.8/threading.py", line 870, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in <lambda>
    target=lambda: self._read_inputs(elements_iterator),
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38225 {created_time:"2023-10-16T08:43:53.069675919+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>

> Task :sdks:python:test-suites:portable:py38:postCommitPy38
> Task :sdks:python:test-suites:dataflow:py38:inferencePostCommitIT

> Task :sdks:python:test-suites:dataflow:py38:postCommitIT
2023-10-16 08:59:52.181614: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.237463: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.250632: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.251360: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:52.301758: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.302469: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:52.309392: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.374666: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.375330: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:52.375470: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.433472: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.438517: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.439256: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:52.496571: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.497259: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:52.510258: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.572700: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.573361: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:52.611872: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.675197: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.675908: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:52.717249: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.781234: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-16 08:59:52.781915: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-16 08:59:53.282834: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-16 08:59:53.333816: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-16 08:59:53.431068: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-16 08:59:53.532516: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-16 08:59:53.611557: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-16 08:59:53.779640: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-16 08:59:53.983810: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-16 08:59:54.195927: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8>: No module named build

> Task :sdks:python:test-suites:dataflow:py38:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 3h 11m 52s
221 actionable tasks: 157 executed, 60 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/cqswakqgk2ohy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python38 #4714

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4714/display/redirect>

Changes:


------------------------------------------
[...truncated 11.05 MB...]
[gw3] linux -- Python 3.8.10 <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8>

args = (['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmp47ojuvcz', ...],)
kwargs = {}

    def check_output(*args, **kwargs):
      if force_shell:
        kwargs['shell'] = True
      try:
>       out = subprocess.check_output(*args, **kwargs)

apache_beam/utils/processes.py:89: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
/usr/lib/python3.8/subprocess.py:415: in check_output
    return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

input = None, capture_output = False, timeout = None, check = True
popenargs = (['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmp47ojuvcz', ...],)
kwargs = {'stdout': -1}, process = <subprocess.Popen object at 0x7f82dc4ad550>
stdout = b'', stderr = None, retcode = 1

    def run(*popenargs,
            input=None, capture_output=False, timeout=None, check=False, **kwargs):
        """Run command with arguments and return a CompletedProcess instance.
    
        The returned instance will have attributes args, returncode, stdout and
        stderr. By default, stdout and stderr are not captured, and those attributes
        will be None. Pass stdout=PIPE and/or stderr=PIPE in order to capture them.
    
        If check is True and the exit code was non-zero, it raises a
        CalledProcessError. The CalledProcessError object will have the return code
        in the returncode attribute, and output & stderr attributes if those streams
        were captured.
    
        If timeout is given, and the process takes too long, a TimeoutExpired
        exception will be raised.
    
        There is an optional argument "input", allowing you to
        pass bytes or a string to the subprocess's stdin.  If you use this argument
        you may not also use the Popen constructor's "stdin" argument, as
        it will be used internally.
    
        By default, all communication is in bytes, and therefore any "input" should
        be bytes, and the stdout and stderr will be bytes. If in text mode, any
        "input" should be a string, and stdout and stderr will be strings decoded
        according to locale encoding, or by "encoding" if set. Text mode is
        triggered by setting any of text, encoding, errors or universal_newlines.
    
        The other arguments are the same as for the Popen constructor.
        """
        if input is not None:
            if kwargs.get('stdin') is not None:
                raise ValueError('stdin and input arguments may not both be used.')
            kwargs['stdin'] = PIPE
    
        if capture_output:
            if kwargs.get('stdout') is not None or kwargs.get('stderr') is not None:
                raise ValueError('stdout and stderr arguments may not be used '
                                 'with capture_output.')
            kwargs['stdout'] = PIPE
            kwargs['stderr'] = PIPE
    
        with Popen(*popenargs, **kwargs) as process:
            try:
                stdout, stderr = process.communicate(input, timeout=timeout)
            except TimeoutExpired as exc:
                process.kill()
                if _mswindows:
                    # Windows accumulates the output in a single blocking
                    # read() call run on child threads, with the timeout
                    # being done in a join() on those threads.  communicate()
                    # _after_ kill() is required to collect that and add it
                    # to the exception.
                    exc.stdout, exc.stderr = process.communicate()
                else:
                    # POSIX _communicate already populated the output so
                    # far into the TimeoutExpired exception.
                    process.wait()
                raise
            except:  # Including KeyboardInterrupt, communicate handled that.
                process.kill()
                # We don't call process.wait() as .__exit__ does that for us.
                raise
            retcode = process.poll()
            if check and retcode:
>               raise CalledProcessError(retcode, process.args,
                                         output=stdout, stderr=stderr)
E               subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmp47ojuvcz', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.

/usr/lib/python3.8/subprocess.py:516: CalledProcessError

During handling of the above exception, another exception occurred:

self = <apache_beam.examples.complete.juliaset.juliaset.juliaset_test_it.JuliaSetTestIT testMethod=test_run_example_with_setup_file>

    def test_run_example_with_setup_file(self):
      pipeline = TestPipeline(is_integration_test=True)
      coordinate_output = FileSystems.join(
          pipeline.get_option('output'),
          'juliaset-{}'.format(str(uuid.uuid4())),
          'coordinates.txt')
      extra_args = {
          'coordinate_output': coordinate_output,
          'grid_size': self.GRID_SIZE,
          'setup_file': os.path.normpath(
              os.path.join(os.path.dirname(__file__), '..', 'setup.py')),
          'on_success_matcher': all_of(PipelineStateMatcher(PipelineState.DONE)),
      }
      args = pipeline.get_full_options_as_args(**extra_args)
    
>     juliaset.run(args)

apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py:56: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/examples/complete/juliaset/juliaset/juliaset.py:116: in run
    (
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:53: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/dataflow/dataflow_runner.py:393: in run_pipeline
    artifacts = environments.python_sdk_dependencies(options)
apache_beam/transforms/environments.py:846: in python_sdk_dependencies
    return stager.Stager.create_job_resources(
apache_beam/runners/portability/stager.py:281: in create_job_resources
    tarball_file = Stager._build_setup_package(
apache_beam/runners/portability/stager.py:796: in _build_setup_package
    processes.check_output(build_setup_args)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

args = (['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmp47ojuvcz', ...],)
kwargs = {}

    def check_output(*args, **kwargs):
      if force_shell:
        kwargs['shell'] = True
      try:
        out = subprocess.check_output(*args, **kwargs)
      except OSError:
        raise RuntimeError("Executable {} not found".format(args[0]))
      except subprocess.CalledProcessError as error:
        if isinstance(args, tuple) and (args[0][2] == "pip"):
          raise RuntimeError( \
            "Full traceback: {} \n Pip install failed for package: {} \
            \n Output from execution of subprocess: {}" \
            .format(traceback.format_exc(), args[0][6], error.output))
        else:
>         raise RuntimeError("Full trace: {}, \
             output of the failed child process {} "\
            .format(traceback.format_exc(), error.output))
E         RuntimeError: Full trace: Traceback (most recent call last):
E           File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/processes.py",> line 89, in check_output
E             out = subprocess.check_output(*args, **kwargs)
E           File "/usr/lib/python3.8/subprocess.py", line 415, in check_output
E             return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
E           File "/usr/lib/python3.8/subprocess.py", line 516, in run
E             raise CalledProcessError(retcode, process.args,
E         subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmp47ojuvcz', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.
E         ,            output of the failed child process b''

apache_beam/utils/processes.py:99: RuntimeError
------------------------------ Captured log call -------------------------------
INFO     apache_beam.runners.portability.stager:stager.py:762 Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', '/tmp/tmpp6_hcxje/tmp_requirements.txt', '--exists-action', 'i', '--no-deps', '--implementation', 'cp', '--abi', 'cp38', '--platform', 'manylinux2014_x86_64']
INFO     apache_beam.runners.portability.stager:stager.py:795 Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmp47ojuvcz', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']>
=============================== warnings summary ===============================
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo?  You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
    @pytest.mark.uses_vertex_ai

apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    dataset_ref = client.dataset(unique_dataset_name, project=project)

apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    table_ref = client.dataset(dataset_id).table(table_id)

apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
    return airline_df[at_top_airports].mean()

apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
=========================== short test summary info ============================
FAILED apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py::JuliaSetTestIT::test_run_example_with_setup_file - RuntimeError: Full trace: Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/processes.py",> line 89, in check_output
    out = subprocess.check_output(*args, **kwargs)
  File "/usr/lib/python3.8/subprocess.py", line 415, in check_output
    return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
  File "/usr/lib/python3.8/subprocess.py", line 516, in run
    raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python3.8',> '-m', 'build', '--sdist', '--outdir', '/tmp/tmp47ojuvcz', '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/complete/juliaset']'> returned non-zero exit status 1.
,            output of the failed child process b''
====== 1 failed, 88 passed, 49 skipped, 17 warnings in 8264.67s (2:17:44) ======

> Task :sdks:python:test-suites:dataflow:py38:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 3h 2m 52s
221 actionable tasks: 157 executed, 60 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/3lye6yiqjib5y

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python38 #4713

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4713/display/redirect>

Changes:


------------------------------------------
[...truncated 9.13 MB...]
  seconds: 1697402621
  nanos: 129973173
}
message: "Creating client data channel for localhost:41083"
instruction_id: "bundle_1"
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/data_plane.py:770"
thread: "Thread-13"

INFO:root:severity: INFO
timestamp {
  seconds: 1697402621
  nanos: 165221452
}
message: "No more requests from control plane"
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker.py:274"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1697402621
  nanos: 165339708
}
message: "SDK Harness waiting for in-flight requests to complete"
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker.py:275"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1697402621
  nanos: 165414810
}
message: "Closing all cached grpc data channels."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/data_plane.py:803"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1697402621
  nanos: 165480136
}
message: "Closing all cached gRPC state handlers."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker.py:904"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1697402621
  nanos: 166231155
}
message: "Done consuming work."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker.py:287"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1697402621
  nanos: 166332960
}
message: "Python sdk harness exiting."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker_main.py:211"
thread: "MainThread"

INFO:apache_beam.runners.portability.local_job_service:Completed job in 32.205605268478394 seconds with state DONE.
INFO:root:Completed job in 32.205605268478394 seconds with state DONE.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE

> Task :sdks:python:test-suites:portable:py38:portableWordCountSparkRunnerBatch
INFO:apache_beam.runners.worker.worker_pool_main:Listening for workers at localhost:37659
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7fbb16f30e50> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7fbb16f30ee0> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7fbb16f32670> ====================
INFO:apache_beam.utils.subprocess_server:Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/runners/spark/3/job-server/build/libs/beam-runners-spark-3-job-server-2.52.0-SNAPSHOT.jar'> '--spark-master-url' 'local[4]' '--artifacts-dir' '/tmp/beam-temp_q5hddg6/artifacts8tnzxrzc' '--job-port' '60629' '--artifact-port' '0' '--expansion-port' '0']
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:55 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:56 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:35545
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:56 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:34307
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:56 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:60629
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:56 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:57 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_bd1a1c5e-44f2-472c-b8ff-c3bb39ff758b.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:57 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_bd1a1c5e-44f2-472c-b8ff-c3bb39ff758b.ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:57 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_bd1a1c5e-44f2-472c-b8ff-c3bb39ff758b.null.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:58 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_bd1a1c5e-44f2-472c-b8ff-c3bb39ff758b.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:58 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-1015204358-50ee2432_5c85f1c5-f027-4013-b905-bfec8cb27b00
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:58 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-1015204358-50ee2432_5c85f1c5-f027-4013-b905-bfec8cb27b00
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
  with Pipeline() as p:
    p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:58 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:43:59 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.util.log: Logging initialized @7143ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_382-8u382-ga-1~20.04.1-b05
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.Server: Started @7263ms
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@7d6444d1{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3bdaaa7d{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@65668a18{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3570f85a{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@356d470d{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@40d4c1{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@322c3543{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5d96fafc{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5640df32{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2832694d{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@18caa9a0{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5476d23f{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4bfb7db9{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6b3da76f{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@771ddc62{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3bb5fd59{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1e69c4f6{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1e99c899{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3ecc7905{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@71112392{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@50146fb6{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6e6abec4{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@411bfbaa{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5bd660f2{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@40c0e066{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4ebed54d{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@34160e9b{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:00 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-1015204358-50ee2432_5c85f1c5-f027-4013-b905-bfec8cb27b00 on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:36383.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:43251.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:39563
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:04 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1015204358-50ee2432_5c85f1c5-f027-4013-b905-bfec8cb27b00: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:04 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1015204358-50ee2432_5c85f1c5-f027-4013-b905-bfec8cb27b00 finished.
INFO:apache_beam.utils.subprocess_server:23/10/15 20:44:04 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@7d6444d1{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread run_worker_1-1:
Traceback (most recent call last):
  File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
    self.run()
  File "/usr/lib/python3.8/threading.py", line 870, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 262, in run
    for work_request in self._control_stub.Control(get_responses()):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
Exception in thread read_state:
Traceback (most recent call last):
  File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
    self.run()
  File "/usr/lib/python3.8/threading.py", line 870, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1035, in pull_responses
    for response in responses:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:39563 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-15T20:44:04.815398034+00:00"}"
>
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:43251 {created_time:"2023-10-15T20:44:04.815408011+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
  File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:36383 {created_time:"2023-10-15T20:44:04.815427434+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
    self.run()
  File "/usr/lib/python3.8/threading.py", line 870, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in <lambda>
    target=lambda: self._read_inputs(elements_iterator),
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:39563 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-15T20:44:04.815398034+00:00"}"
>

> Task :sdks:python:test-suites:portable:py38:postCommitPy38

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:installGcpTest'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 28m 43s
217 actionable tasks: 153 executed, 60 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/zyr4jpw7djebk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python38 #4712

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4712/display/redirect>

Changes:


------------------------------------------
[...truncated 85.47 KB...]
go: added github.com/golang/protobuf v1.5.3
go: added golang.org/x/net v0.15.0
go: added golang.org/x/sys v0.12.0
go: added golang.org/x/text v0.13.0
go: added google.golang.org/genproto/googleapis/rpc v0.0.0-20230913181813-007df8e322eb
go: added google.golang.org/grpc v1.58.1
go: added google.golang.org/protobuf v1.31.0
+ go-licenses save github.com/apache/beam/sdks/v2/java/container --save_path=/output/licenses

> Task :sdks:java:container:java8:copyJavaThirdPartyLicenses

> Task :sdks:java:container:downloadCloudProfilerAgent
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed
  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0  0 4183k    0 22141    0     0   450k      0  0:00:09 --:--:--  0:00:09  441kversion.txt
NOTICES
profiler_java_agent.so
100 4183k  100 4183k    0     0  25.8M      0 --:--:-- --:--:-- --:--:-- 25.6M

> Task :sdks:java:container:goPrepare UP-TO-DATE

> Task :sdks:java:container:goBuild
/home/jenkins/go/bin/go1.21.2 build -o ./build/target/linux_amd64/boot boot.go boot_test.go pathingjar.go pathingjar_test.go

> Task :sdks:java:container:java8:copySdkHarnessLauncher
> Task :runners:flink:1.16:job-server:shadowJar FROM-CACHE

> Task :release:go-licenses:java:dockerRun
W1015 14:16:27.272640      42 library.go:101] "golang.org/x/sys/unix" contains non-Go code that can't be inspected for further dependencies:
/go/pkg/mod/golang.org/x/sys@v0.12.0/unix/asm_linux_amd64.s

> Task :sdks:python:test-suites:direct:py38:hdfsIntegrationTest
Removing intermediate container b09a9f61020e
 ---> 098860a0d00a
Step 8/8 : CMD cd sdks/python && tox -e hdfs_integration_test
 ---> Running in 47c30208ca55
Removing intermediate container 47c30208ca55
 ---> 7c858ec3c252

> Task :release:go-licenses:py:createLicenses
> Task :sdks:python:container:py38:copyGolangLicenses

> Task :release:go-licenses:java:dockerRun
+ go-licenses csv github.com/apache/beam/sdks/v2/java/container
+ tee /output/licenses/list.csv
W1015 14:16:31.446283      86 library.go:101] "golang.org/x/sys/unix" contains non-Go code that can't be inspected for further dependencies:
/go/pkg/mod/golang.org/x/sys@v0.12.0/unix/asm_linux_amd64.s
+ chmod -R a+w /output/licenses

> Task :release:go-licenses:java:createLicenses

> Task :sdks:python:test-suites:direct:py38:hdfsIntegrationTest

real	0m41.539s
user	0m0.828s
sys	0m0.290s
+ docker-compose -p hdfs_IT-jenkins-beam_PostCommit_Python38-4712 --no-ansi up --exit-code-from test --abort-on-container-exit --force-recreate
--no-ansi option is deprecated and will be removed in future versions. Use `--ansi never` instead.
Creating network "hdfs_it-jenkins-beam_postcommit_python38-4712_test_net" with the default driver
Creating hdfs_it-jenkins-beam_postcommit_python38-4712_namenode_1 ... 
Creating hdfs_it-jenkins-beam_postcommit_python38-4712_namenode_1 ... done
Creating hdfs_it-jenkins-beam_postcommit_python38-4712_datanode_1 ... 
Creating hdfs_it-jenkins-beam_postcommit_python38-4712_datanode_1 ... done
Creating hdfs_it-jenkins-beam_postcommit_python38-4712_test_1     ... 
Creating hdfs_it-jenkins-beam_postcommit_python38-4712_test_1     ... done

> Task :sdks:java:container:java8:copyGolangLicenses
> Task :sdks:java:container:java8:dockerPrepare

> Task :sdks:java:container:java8:docker
#1 [internal] load build definition from Dockerfile
#1 transferring dockerfile: 2.58kB done
#1 DONE 0.0s

#2 [internal] load .dockerignore
#2 transferring context: 2B done
#2 DONE 0.0s

#3 [internal] load metadata for docker.io/library/eclipse-temurin:8
#3 DONE 0.3s

#4 [ 1/16] FROM docker.io/library/eclipse-temurin:8@sha256:5840452260db162c2dd31033cb55575745861634b1be75a1d8cff559b6c8b942
#4 resolve docker.io/library/eclipse-temurin:8@sha256:5840452260db162c2dd31033cb55575745861634b1be75a1d8cff559b6c8b942 0.0s done
#4 CACHED

#5 [internal] load build context
#5 transferring context: 47.28MB 0.4s done
#5 DONE 0.4s

#6 [ 2/16] ADD target/slf4j-api.jar /opt/apache/beam/jars/
#6 DONE 1.1s

#7 [ 3/16] ADD target/slf4j-jdk14.jar /opt/apache/beam/jars/
#7 DONE 0.1s

#8 [ 4/16] ADD target/jcl-over-slf4j.jar /opt/apache/beam/jars/
#8 DONE 0.1s

#9 [ 5/16] ADD target/log4j-over-slf4j.jar /opt/apache/beam/jars/
#9 DONE 0.1s

#10 [ 6/16] ADD target/log4j-to-slf4j.jar /opt/apache/beam/jars/
#10 DONE 0.1s

#11 [ 7/16] ADD target/beam-sdks-java-harness.jar /opt/apache/beam/jars/
#11 DONE 0.1s

#12 [ 8/16] COPY target/jamm.jar target/open-module-agent*.jar /opt/apache/beam/jars/
#12 DONE 0.1s

#13 [ 9/16] COPY target/linux_amd64/boot /opt/apache/beam/
#13 DONE 0.2s

#14 [10/16] COPY target/LICENSE /opt/apache/beam/
#14 DONE 0.1s

#15 [11/16] COPY target/NOTICE /opt/apache/beam/
#15 DONE 0.1s

#16 [12/16] ADD target/third_party_licenses /opt/apache/beam/third_party_licenses/
#16 DONE 0.2s

#17 [13/16] COPY target/LICENSE target/options/* /opt/apache/beam/options/
#17 DONE 0.1s

#18 [14/16] COPY target/go-licenses/* /opt/apache/beam/third_party_licenses/golang/
#18 DONE 0.1s

#19 [15/16] RUN if [ "true" = "false" ] ; then     rm -rf /opt/apache/beam/third_party_licenses ;    fi
#19 DONE 0.4s

#20 [16/16] COPY target/profiler/* /opt/google_cloud_profiler/
#20 DONE 0.1s

#21 exporting to image
#21 exporting layers
#21 exporting layers 0.2s done
#21 writing image sha256:a5e7b050eb65d5fb8af96d785184a87ff488563cf63a2eccb3af79ddf3feadfe done
#21 naming to docker.io/apache/beam_java8_sdk:2.52.0.dev done
#21 DONE 0.2s

> Task :sdks:python:test-suites:direct:py38:hdfsIntegrationTest
Stopping hdfs_it-jenkins-beam_postcommit_python38-4712_datanode_1 ... 
Stopping hdfs_it-jenkins-beam_postcommit_python38-4712_namenode_1 ... 
Stopping hdfs_it-jenkins-beam_postcommit_python38-4712_datanode_1 ... done
Stopping hdfs_it-jenkins-beam_postcommit_python38-4712_namenode_1 ... done

real	6m34.816s
user	0m1.434s
sys	0m0.181s
+ finally
+ docker-compose -p hdfs_IT-jenkins-beam_PostCommit_Python38-4712 --no-ansi down
--no-ansi option is deprecated and will be removed in future versions. Use `--ansi never` instead.
Removing hdfs_it-jenkins-beam_postcommit_python38-4712_test_1     ... 
Removing hdfs_it-jenkins-beam_postcommit_python38-4712_datanode_1 ... 
Removing hdfs_it-jenkins-beam_postcommit_python38-4712_namenode_1 ... 
Removing hdfs_it-jenkins-beam_postcommit_python38-4712_namenode_1 ... done
Removing hdfs_it-jenkins-beam_postcommit_python38-4712_datanode_1 ... done
Removing hdfs_it-jenkins-beam_postcommit_python38-4712_test_1     ... done
Removing network hdfs_it-jenkins-beam_postcommit_python38-4712_test_net

real	0m1.298s
user	0m0.543s
sys	0m0.122s

> Task :sdks:python:test-suites:direct:py38:setupVirtualenv
Collecting pip
  Using cached pip-23.2.1-py3-none-any.whl (2.1 MB)
Installing collected packages: pip
  Attempting uninstall: pip
    Found existing installation: pip 20.0.2
    Uninstalling pip-20.0.2:
      Successfully uninstalled pip-20.0.2
Successfully installed pip-23.2.1
Collecting tox
  Obtaining dependency information for tox from https://files.pythonhosted.org/packages/f5/f9/963052e8b825645c54262dce7b7c88691505e3b9ee10a3e3667711eaaf21/tox-4.11.3-py3-none-any.whl.metadata
  Using cached tox-4.11.3-py3-none-any.whl.metadata (5.0 kB)
Collecting cachetools>=5.3.1 (from tox)
  Obtaining dependency information for cachetools>=5.3.1 from https://files.pythonhosted.org/packages/a9/c9/c8a7710f2cedcb1db9224fdd4d8307c9e48cbddc46c18b515fefc0f1abbe/cachetools-5.3.1-py3-none-any.whl.metadata
  Using cached cachetools-5.3.1-py3-none-any.whl.metadata (5.2 kB)
Collecting chardet>=5.2 (from tox)
  Obtaining dependency information for chardet>=5.2 from https://files.pythonhosted.org/packages/38/6f/f5fbc992a329ee4e0f288c1fe0e2ad9485ed064cac731ed2fe47dcc38cbf/chardet-5.2.0-py3-none-any.whl.metadata
  Using cached chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)
Collecting colorama>=0.4.6 (from tox)
  Using cached colorama-0.4.6-py2.py3-none-any.whl (25 kB)
Collecting filelock>=3.12.3 (from tox)
  Obtaining dependency information for filelock>=3.12.3 from https://files.pythonhosted.org/packages/5e/5d/97afbafd9d584ff1b45fcb354a479a3609bd97f912f8f1f6c563cb1fae21/filelock-3.12.4-py3-none-any.whl.metadata
  Using cached filelock-3.12.4-py3-none-any.whl.metadata (2.8 kB)
Collecting packaging>=23.1 (from tox)
  Obtaining dependency information for packaging>=23.1 from https://files.pythonhosted.org/packages/ec/1a/610693ac4ee14fcdf2d9bf3c493370e4f2ef7ae2e19217d7a237ff42367d/packaging-23.2-py3-none-any.whl.metadata
  Using cached packaging-23.2-py3-none-any.whl.metadata (3.2 kB)
Collecting platformdirs>=3.10 (from tox)
  Obtaining dependency information for platformdirs>=3.10 from https://files.pythonhosted.org/packages/56/29/3ec311dc18804409ecf0d2b09caa976f3ae6215559306b5b530004e11156/platformdirs-3.11.0-py3-none-any.whl.metadata
  Using cached platformdirs-3.11.0-py3-none-any.whl.metadata (11 kB)
Collecting pluggy>=1.3 (from tox)
  Obtaining dependency information for pluggy>=1.3 from https://files.pythonhosted.org/packages/05/b8/42ed91898d4784546c5f06c60506400548db3f7a4b3fb441cba4e5c17952/pluggy-1.3.0-py3-none-any.whl.metadata
  Using cached pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)
Collecting pyproject-api>=1.6.1 (from tox)
  Obtaining dependency information for pyproject-api>=1.6.1 from https://files.pythonhosted.org/packages/cf/b4/39eea50542e50e93876ebc09c4349a9c9eee9f6b9c9d30f88c7dc5433db8/pyproject_api-1.6.1-py3-none-any.whl.metadata
  Using cached pyproject_api-1.6.1-py3-none-any.whl.metadata (2.8 kB)
Collecting tomli>=2.0.1 (from tox)
  Using cached tomli-2.0.1-py3-none-any.whl (12 kB)
Collecting virtualenv>=20.24.3 (from tox)
  Obtaining dependency information for virtualenv>=20.24.3 from https://files.pythonhosted.org/packages/4e/8b/f0d3a468c0186c603217a6656ea4f49259630e8ed99558501d92f6ff7dc3/virtualenv-20.24.5-py3-none-any.whl.metadata
  Using cached virtualenv-20.24.5-py3-none-any.whl.metadata (4.5 kB)
Collecting distlib<1,>=0.3.7 (from virtualenv>=20.24.3->tox)
  Obtaining dependency information for distlib<1,>=0.3.7 from https://files.pythonhosted.org/packages/43/a0/9ba967fdbd55293bacfc1507f58e316f740a3b231fc00e3d86dc39bc185a/distlib-0.3.7-py2.py3-none-any.whl.metadata
  Using cached distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)
Using cached tox-4.11.3-py3-none-any.whl (153 kB)
Using cached cachetools-5.3.1-py3-none-any.whl (9.3 kB)
Using cached chardet-5.2.0-py3-none-any.whl (199 kB)
Using cached filelock-3.12.4-py3-none-any.whl (11 kB)
Using cached packaging-23.2-py3-none-any.whl (53 kB)
Using cached platformdirs-3.11.0-py3-none-any.whl (17 kB)
Using cached pluggy-1.3.0-py3-none-any.whl (18 kB)
Using cached pyproject_api-1.6.1-py3-none-any.whl (12 kB)
Using cached virtualenv-20.24.5-py3-none-any.whl (3.7 MB)
Using cached distlib-0.3.7-py2.py3-none-any.whl (468 kB)
Installing collected packages: distlib, tomli, pluggy, platformdirs, packaging, filelock, colorama, chardet, cachetools, virtualenv, pyproject-api, tox
Successfully installed cachetools-5.3.1 chardet-5.2.0 colorama-0.4.6 distlib-0.3.7 filelock-3.12.4 packaging-23.2 platformdirs-3.11.0 pluggy-1.3.0 pyproject-api-1.6.1 tomli-2.0.1 tox-4.11.3 virtualenv-20.24.5

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/build.gradle'> line: 48

* What went wrong:
Execution failed for task ':sdks:python:sdist'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 7m 50s
200 actionable tasks: 136 executed, 60 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/3ogec5pvvd7dw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org