You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2023/10/06 05:09:12 UTC
Build failed in Jenkins: beam_PostCommit_Python38 #4674
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4674/display/redirect>
Changes:
------------------------------------------
[...truncated 15.70 MB...]
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_kms
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw0] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_1M_python
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw4] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw6] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_1M_python
apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python
[gw4] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw5] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw0] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw6] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw6] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw6] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw6] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw5] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw1] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw6] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw6] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw2] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw2] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw2] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw2] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw0] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw0] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw0] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw7] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 7864.82s (2:11:04)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 2h 53m 40s
215 actionable tasks: 154 executed, 57 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/fqzvdci5nycb2
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Jenkins build is back to normal : beam_PostCommit_Python38 #4691
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4691/display/redirect>
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4690
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4690/display/redirect>
Changes:
------------------------------------------
[...truncated 11.35 MB...]
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:18 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1010023212-df68fd9d_347faf85-cc28-45f4-b3e4-95282df3fdda: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:18 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1010023212-df68fd9d_347faf85-cc28-45f4-b3e4-95282df3fdda finished.
INFO:apache_beam.utils.subprocess_server:23/10/10 02:32:18 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@4d6cc072{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread read_state:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
Exception in thread run_worker_1-1:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:42863 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-10T02:32:18.78914216+00:00"}"
>
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 262, in run
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
for work_request in self._control_stub.Control(get_responses()):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1035, in pull_responses
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in <lambda>
target=lambda: self._read_inputs(elements_iterator),
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
for response in responses:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38019 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-10T02:32:18.789145307+00:00"}"
>
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:44743 {created_time:"2023-10-10T02:32:18.789142258+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:42863 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-10T02:32:18.78914216+00:00"}"
>
> Task :sdks:python:test-suites:portable:py38:postCommitPy38
> Task :sdks:python:test-suites:direct:py38:postCommitIT
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_temp_table_append_schema_update_2
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/1398941892/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/datastore/v1new/datastoreio.py>:250: UserWarning: Detected filter using positional arguments. Prefer using the 'filter' keyword argument instead.
query.add_filter('kind_name', '=', kind_name)
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/datastore/v1new/datastoreio.py>:251: UserWarning: Detected filter using positional arguments. Prefer using the 'filter' keyword argument instead.
query.add_filter('timestamp', '=', latest_timestamp)
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/1398941892/lib/python3.8/site-packages/google/cloud/datastore/query.py>:234: UserWarning: Detected filter using positional arguments. Prefer using the 'filter' keyword argument instead.
self.add_filter(property_name, operator, value)
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-direct-py38.xml> -
[33m================= [32m38 passed[0m, [33m[1m14 warnings[0m[33m in 143.97s (0:02:23)[0m[33m ==================[0m
> Task :sdks:python:test-suites:dataflow:py38:initializeForDataflowJob
> Task :sdks:python:test-suites:dataflow:py38:vertexAIInferenceTest
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
azure-core 1.29.4 requires typing-extensions>=4.6.0, but you have typing-extensions 4.5.0 which is incompatible.
2023-10-10 02:33:50.815967: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.816293: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.816332: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.826373: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.837143: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.879041: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.879040: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.879040: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.879786: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:50.879787: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:50.879787: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:50.886220: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.888518: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.889216: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:50.899588: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.900307: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:50.950353: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:50.951064: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:50.965970: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:51.030562: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:51.031317: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:51.038334: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:51.102092: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:33:51.102905: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:33:51.959706: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:33:52.017908: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:33:52.019915: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:33:52.044466: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:33:52.051599: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:33:52.055119: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:33:52.104615: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:33:52.214926: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
> Task :sdks:python:test-suites:dataflow:py38:inferencePostCommitIT
> Task :sdks:python:test-suites:dataflow:py38:postCommitIT
2023-10-10 02:57:00.240867: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.294688: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.306354: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.307099: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:00.347222: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.357940: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.358673: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:00.411815: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.412521: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:00.484244: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.547452: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.548193: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:00.556599: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.619235: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.619961: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:00.645940: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.709635: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.710382: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:00.749871: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.794551: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.813013: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.813779: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:00.857350: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-10 02:57:00.858043: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-10 02:57:01.337112: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:57:01.386340: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:57:01.448686: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:57:01.719663: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:57:01.830050: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:57:01.931142: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:57:02.167459: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-10 02:57:02.217386: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md
warning: check: missing required meta-data: url
warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) must be supplied
> Task :sdks:python:test-suites:dataflow:py38:postCommitIT FAILED
FAILURE: Build failed with an exception.
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 1m 24s
221 actionable tasks: 157 executed, 60 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/gz3oqc7fsl3va
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4689
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4689/display/redirect>
Changes:
------------------------------------------
[...truncated 11.21 MB...]
}
message: "Closing all cached grpc data channels."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/data_plane.py:803"
thread: "MainThread"
INFO:root:severity: INFO
timestamp {
seconds: 1696883867
nanos: 487160682
}
message: "Closing all cached gRPC state handlers."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker.py:904"
thread: "MainThread"
INFO:root:severity: INFO
timestamp {
seconds: 1696883867
nanos: 487907171
}
message: "Done consuming work."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker.py:287"
thread: "MainThread"
INFO:root:severity: INFO
timestamp {
seconds: 1696883867
nanos: 488008260
}
message: "Python sdk harness exiting."
log_location: "/usr/local/lib/python3.8/site-packages/apache_beam/runners/worker/sdk_worker_main.py:211"
thread: "MainThread"
INFO:apache_beam.runners.portability.local_job_service:Completed job in 30.487509965896606 seconds with state DONE.
INFO:root:Completed job in 30.487509965896606 seconds with state DONE.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
> Task :sdks:python:test-suites:portable:py38:portableWordCountSparkRunnerBatch
INFO:apache_beam.runners.worker.worker_pool_main:Listening for workers at localhost:39751
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7f853c285af0> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f853c285b80> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7f853c284310> ====================
INFO:apache_beam.utils.subprocess_server:Starting service with ['java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/runners/spark/3/job-server/build/libs/beam-runners-spark-3-job-server-2.52.0-SNAPSHOT.jar'> '--spark-master-url' 'local[4]' '--artifacts-dir' '/tmp/beam-tempfzqvx6ih/artifacts9jf3lnfq' '--job-port' '56829' '--artifact-port' '0' '--expansion-port' '0']
INFO:apache_beam.utils.subprocess_server:23/10/09 20:37:59 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:00 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:41025
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:00 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:45901
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:00 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:56829
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:00 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
WARNING:root:Waiting for grpc channel to be ready at localhost:56829.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:01 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_09d0b9ed-6ff0-4c2f-a4d8-bee6485b9968.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:01 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_09d0b9ed-6ff0-4c2f-a4d8-bee6485b9968.ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:01 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_09d0b9ed-6ff0-4c2f-a4d8-bee6485b9968.null.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:01 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_09d0b9ed-6ff0-4c2f-a4d8-bee6485b9968.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:02 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-1009203801-457cc365_dee9e841-a820-4e43-99be-3ad40662c531
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:02 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-1009203801-457cc365_dee9e841-a820-4e43-99be-3ad40662c531
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
with Pipeline() as p:
p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:02 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:03 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:04 INFO org.sparkproject.jetty.util.log: Logging initialized @8687ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:04 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_382-8u382-ga-1~20.04.1-b05
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.Server: Started @8833ms
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@61516bb4{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4d14cd1d{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@157b18fa{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@73b3bdc4{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6203f6fa{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@43d95d55{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@378e5144{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@33de4388{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5d90f4bb{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@13f1a5fe{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4c3dacd9{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2bcaa723{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@9dd7170{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@136f08a0{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@63b057b6{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@77209dad{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@30eebf77{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@524474f0{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@542dfb1e{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@56723c04{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@13e59978{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7289c314{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@38d6a8e4{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1256e852{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5ce7a52e{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3557a132{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@8973516{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:05 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-1009203801-457cc365_dee9e841-a820-4e43-99be-3ad40662c531 on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:46817.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:33003.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:46057
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:11 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:12 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:13 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1009203801-457cc365_dee9e841-a820-4e43-99be-3ad40662c531: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:13 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1009203801-457cc365_dee9e841-a820-4e43-99be-3ad40662c531 finished.
INFO:apache_beam.utils.subprocess_server:23/10/09 20:38:13 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@61516bb4{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread run_worker_1-1:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
Exception in thread read_state:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "recvmsg:Connection reset by peer"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:46057 {created_time:"2023-10-09T20:38:13.824315616+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in <lambda>
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 262, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1035, in pull_responses
for response in responses:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
for work_request in self._control_stub.Control(get_responses()):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
target=lambda: self._read_inputs(elements_iterator),
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "recvmsg:Connection reset by peer"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:33003 {grpc_message:"recvmsg:Connection reset by peer", grpc_status:14, created_time:"2023-10-09T20:38:13.824358592+00:00"}"
>
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "recvmsg:Connection reset by peer"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:46057 {created_time:"2023-10-09T20:38:13.824315616+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "recvmsg:Connection reset by peer"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:46817 {created_time:"2023-10-09T20:38:13.824392507+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
> Task :sdks:python:test-suites:portable:py38:postCommitPy38
> Task :sdks:python:test-suites:dataflow:py38:postCommitIT
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md
warning: check: missing required meta-data: url
warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) must be supplied
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 420
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:tensorRTtests'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 448
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:vertexAIInferenceTest'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 5m 51s
221 actionable tasks: 160 executed, 57 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/djb6s3xm7ezt2
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4688
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4688/display/redirect?page=changes>
Changes:
[noreply] Bump org.nosphere.apache.rat from 0.8.0 to 0.8.1 (#28675)
------------------------------------------
[...truncated 16.06 MB...]
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
[gw6] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw4] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_multiple_destinations_transform
[gw6] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw4] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw4] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_multiple_destinations_transform
apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform
[gw5] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw6] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw7] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw4] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw4] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
[gw6] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw7] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw4] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw4] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw1] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw5] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw5] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw6] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8954.62s (2:29:14)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 9m 26s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/zw6agnsdte5bs
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4687
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4687/display/redirect>
Changes:
------------------------------------------
[...truncated 11.97 MB...]
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2f55b7b5{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@40628eeb{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7c3115d5{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@8a7a163{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5e0b68e5{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2edf049{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@72b6989f{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/09 08:32:59 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-1009083257-2335a60e_a92afe0b-fd33-43df-a488-1ff2d769e7d0 on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:44235.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:01 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:01 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:40623.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:36863
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:01 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:01 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1009083257-2335a60e_a92afe0b-fd33-43df-a488-1ff2d769e7d0: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1009083257-2335a60e_a92afe0b-fd33-43df-a488-1ff2d769e7d0 finished.
INFO:apache_beam.utils.subprocess_server:23/10/09 08:33:02 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@1a6cd394{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread run_worker_1-1:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
Exception in thread read_state:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 262, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1035, in pull_responses
for response in responses:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
for work_request in self._control_stub.Control(get_responses()):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:40623 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-09T08:33:03.412601205+00:00"}"
>
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "recvmsg:Connection reset by peer"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:44235 {grpc_message:"recvmsg:Connection reset by peer", grpc_status:14, created_time:"2023-10-09T08:33:03.412613943+00:00"}"
>
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:36863 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-09T08:33:03.412587515+00:00"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in <lambda>
target=lambda: self._read_inputs(elements_iterator),
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:36863 {grpc_message:"Socket closed", grpc_status:14, created_time:"2023-10-09T08:33:03.412587515+00:00"}"
>
> Task :sdks:python:test-suites:direct:py38:postCommitIT FAILED
> Task :sdks:python:test-suites:dataflow:py38:initializeForDataflowJob
> Task :sdks:python:test-suites:dataflow:py38:vertexAIInferenceTest
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
azure-core 1.29.4 requires typing-extensions>=4.6.0, but you have typing-extensions 4.5.0 which is incompatible.
2023-10-09 08:34:24.366656: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.402833: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.440118: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.441047: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:24.481618: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.482585: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:24.510815: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.530868: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.566827: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.601191: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.602282: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:24.608074: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.625573: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.626453: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:24.638533: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.649468: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.659846: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.660725: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:24.707455: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.708349: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:24.715026: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.715750: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:24.739313: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:34:24.740245: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:34:25.716068: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:34:25.912816: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:34:25.997240: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:34:26.034273: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:34:26.036528: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:34:26.163082: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:34:26.661863: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:34:26.705708: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
> Task :sdks:python:test-suites:dataflow:py38:inferencePostCommitIT
> Task :sdks:python:test-suites:dataflow:py38:postCommitIT
2023-10-09 08:56:37.200328: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.232943: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.315080: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.316145: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:37.351077: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.352340: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:37.580466: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.671122: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.672114: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:37.715625: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.737158: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.817876: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.818740: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:37.829181: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.830359: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:37.840571: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.945085: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.947088: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:37.947712: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:38.059272: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:38.060199: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:38.383970: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:38.513428: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-09 08:56:38.514522: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-09 08:56:39.067791: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:56:39.167641: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:56:39.573288: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:56:39.579966: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:56:39.881963: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:56:39.892154: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:56:40.010554: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-09 08:56:40.521254: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md
warning: check: missing required meta-data: url
warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) must be supplied
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 11m 25s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/xq3ukcjx6bioc
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4686
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4686/display/redirect>
Changes:
------------------------------------------
[...truncated 16.37 MB...]
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw5] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw5] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw1] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw1] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
[gw6] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw0] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw1] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw1] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
[gw6] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw6] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw6] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw5] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw5] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8406.84s (2:20:06)[0m[33m ===========[0m
FAILURE: Build completed with 3 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
3: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 448
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:vertexAIInferenceTest'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 2h 47m 25s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/wmx2kjq7rxxmu
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4685
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4685/display/redirect>
Changes:
------------------------------------------
[...truncated 16.21 MB...]
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw3] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw0] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw3] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw1] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw0] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw5] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw5] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw1] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw1] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw2] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw4] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw4] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8200.68s (2:16:40)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 2h 56m 22s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/d3ifpf4jhvtee
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4684
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4684/display/redirect>
Changes:
------------------------------------------
[...truncated 15.98 MB...]
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw4] [32mPASSED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_kms
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw4] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw6] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_multiple_destinations_transform
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw4] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw5] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw6] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw1] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_multiple_destinations_transform
apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform
[gw4] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw5] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw5] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw5] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw5] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw3] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw6] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
[gw4] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw4] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw0] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw6] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw1] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw6] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw1] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw6] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw1] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw1] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8199.25s (2:16:39)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 2h 55m 39s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/tjaxfxu43usgg
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4683
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4683/display/redirect>
Changes:
------------------------------------------
[...truncated 16.29 MB...]
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw4] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
[gw4] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw4] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw5] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw3] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw3] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw3] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw3] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw4] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw0] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw3] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw3] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw4] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw5] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw5] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw4] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw1] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw1] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw4] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw1] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw6] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8506.21s (2:21:46)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 52s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/ykuxbddcfnmbc
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4682
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4682/display/redirect>
Changes:
------------------------------------------
[...truncated 16.17 MB...]
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_batch_rewrite_token
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw3] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
[gw4] [32mPASSED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw6] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw6] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw6] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw6] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
[gw6] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw6] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw0] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw5] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw4] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw1] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw1] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw7] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw1] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw1] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 9845.40s (2:44:05)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 24m 18s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/4docy2uk7yfdg
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4681
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4681/display/redirect>
Changes:
------------------------------------------
[...truncated 16.16 MB...]
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_kms
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw3] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
[gw4] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw0] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw3] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw1] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw0] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw0] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw5] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw5] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw4] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw4] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw6] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw4] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw4] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw4] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8313.79s (2:18:33)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 2h 58m 51s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/m5n4r4anybo5k
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4680
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4680/display/redirect?page=changes>
Changes:
[noreply] Update executionResult (#28884)
------------------------------------------
[...truncated 16.31 MB...]
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_rename
apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_rename_error
[gw3] [32mPASSED[0m apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_rename_error
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw5] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw4] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw1] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw2] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw4] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw5] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw3] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw3] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw3] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw5] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw1] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw1] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw0] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw5] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw6] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw5] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8422.89s (2:20:22)[0m[33m ===========[0m
FAILURE: Build completed with 4 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 420
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:tensorRTtests'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 448
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:vertexAIInferenceTest'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
3: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
4: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 2h 35m 34s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/f2na2ibqsqadi
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4679
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4679/display/redirect>
Changes:
------------------------------------------
[...truncated 16.17 MB...]
[gw4] [32mPASSED[0m apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy_rewrite_token
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw4] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
[gw6] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
[gw6] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_direct_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
[gw3] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw6] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw6] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw4] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw0] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw7] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw6] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw4] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw4] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw7] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw3] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw7] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw3] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw5] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw3] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 8773.94s (2:26:13)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 5m 49s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/rllaobsiry5qy
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4678
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4678/display/redirect>
Changes:
------------------------------------------
[...truncated 16.34 MB...]
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_1M_python
apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_export_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryStreamingInsertTransformIntegrationTests::test_value_provider_transform
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
[gw6] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
[gw3] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw6] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
[gw1] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw1] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw1] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw1] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw1] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw2] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw2] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw0] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw1] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw1] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
[gw4] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw6] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw6] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 9919.39s (2:45:19)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 26m 26s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/z2u7soh3ypqra
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4677
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4677/display/redirect>
Changes:
------------------------------------------
[...truncated 9.34 MB...]
WARNING:root:Waiting for grpc channel to be ready at localhost:57803.
WARNING:root:Waiting for grpc channel to be ready at localhost:57803.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:14 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
WARNING:root:Waiting for grpc channel to be ready at localhost:57803.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:16 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:45305
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:16 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:39947
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:16 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:57803
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:16 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:18 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_955ee876-38a5-4910-859e-a2a1646ca71c.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:18 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_955ee876-38a5-4910-859e-a2a1646ca71c.ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:18 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_955ee876-38a5-4910-859e-a2a1646ca71c.null.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:18 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_955ee876-38a5-4910-859e-a2a1646ca71c.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:19 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-1006211319-e5e7dffe_d9a7549d-ac23-4b73-80c7-4f23e7192cae
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:19 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-1006211319-e5e7dffe_d9a7549d-ac23-4b73-80c7-4f23e7192cae
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
with Pipeline() as p:
p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:20 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:22 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.util.log: Logging initialized @17922ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_382-8u382-ga-1~20.04.1-b05
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.Server: Started @18404ms
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@63d0acf{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@65dfc95{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4c3e942a{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4757abc{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@42548b8d{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@dfde818{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@9ad6ee8{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5b56c88c{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@54cc114c{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@71df4640{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@61586861{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@25929326{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@a9f0b3{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3e079ff7{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@57ad2f1a{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@33b87b1c{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6630fce5{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:25 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3251541a{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5552f173{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@22ab7d41{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@73e1896b{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@ed1199{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@11b2f3e9{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@55afce7c{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2cf410e2{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@61556a3f{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@26232fca{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:26 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:27 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:27 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-1006211319-e5e7dffe_d9a7549d-ac23-4b73-80c7-4f23e7192cae on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 0
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:41033.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:31 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:31 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:39735.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:46563
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:31 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:31 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:33 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:34 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1006211319-e5e7dffe_d9a7549d-ac23-4b73-80c7-4f23e7192cae: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.10 seconds.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:34 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-1006211319-e5e7dffe_d9a7549d-ac23-4b73-80c7-4f23e7192cae finished.
INFO:apache_beam.utils.subprocess_server:23/10/06 21:13:34 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@63d0acf{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread read_state:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1035, in pull_responses
for response in responses:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:39735 {created_time:"2023-10-06T21:13:35.833132268+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
Exception in thread run_worker_1-1:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 262, in run
for work_request in self._control_stub.Control(get_responses()):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:41033 {created_time:"2023-10-06T21:13:35.834659764+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:46563 {created_time:"2023-10-06T21:13:35.832834048+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in <lambda>
target=lambda: self._read_inputs(elements_iterator),
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 652, in _read_inputs
for elements in elements_iterator:
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 541, in __next__
return self._next()
File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/2022703442/lib/python3.8/site-packages/grpc/_channel.py",> line 967, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:46563 {created_time:"2023-10-06T21:13:35.832834048+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
> Task :sdks:python:test-suites:dataflow:py38:inferencePostCommitIT
> Task :sdks:python:test-suites:dataflow:py38:postCommitIT
2023-10-06 21:32:35.907652: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:35.963486: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:35.973717: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:35.974447: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:36.029723: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.030494: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:36.081993: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.146408: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.147140: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:36.176414: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.208186: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.239728: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.240443: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:36.268723: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.272266: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.272989: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:36.331990: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.332703: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:36.401841: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.466128: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.466857: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:36.567465: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.632446: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.
2023-10-06 21:32:36.633170: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-10-06 21:32:37.015867: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-06 21:32:37.072575: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-06 21:32:37.250953: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-06 21:32:37.365924: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-06 21:32:37.433894: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-06 21:32:37.620576: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-06 21:32:37.733220: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
2023-10-06 21:32:38.059696: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md
warning: check: missing required meta-data: url
warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) must be supplied
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 10m 58s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/twta3hx4hddlq
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4676
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4676/display/redirect>
Changes:
------------------------------------------
[...truncated 16.50 MB...]
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw2] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw7] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
[gw1] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw3] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw7] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw7] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw7] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw7] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw5] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw4] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
[gw7] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw7] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw3] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw3] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw0] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw1] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw1] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw1] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw0] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw6] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw6] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m=========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 9217.63s (2:33:37)[0m[33m ===========[0m
FAILURE: Build completed with 3 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
3: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 448
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:vertexAIInferenceTest'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 1m 21s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/7ox53hgsf33xe
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_Python38 #4675
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/4675/display/redirect?page=changes>
Changes:
[noreply] Bump golang.org/x/net from 0.15.0 to 0.16.0 in /sdks (#28858)
------------------------------------------
[...truncated 16.12 MB...]
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_file_loads_write
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_multiple_destinations_transform
apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_new_types_avro
apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
[gw6] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
[gw4] [32mPASSED[0m apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_query_read
apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
[gw1] [32mPASSED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_one_job_fails_all_jobs_fail
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_errors_reporting
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
[gw3] [32mPASSED[0m apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
[gw5] [32mPASSED[0m apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
[gw5] [33mSKIPPED[0m apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
[gw5] [33mSKIPPED[0m apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
[gw5] [33mSKIPPED[0m apache_beam/ml/inference/pytorch_inference_it_test.py::PyTorchInference::test_torch_run_inference_imagenet_mobilenetv2
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
[gw4] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
[gw1] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_insert_non_transient_api_call_error
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
[gw6] [32mPASSED[0m apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
[gw4] [32mPASSED[0m apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs
apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/base_test.py::RunInferenceBaseTest::test_run_inference_with_side_inputin_streaming
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_imagenet_image_segmentation
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_classification_large_model
apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
[gw4] [33mSKIPPED[0m apache_beam/ml/inference/tensorflow_inference_it_test.py::TensorflowInference::test_tf_mnist_with_weights_classification
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
[gw4] [33mSKIPPED[0m apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
[gw4] [33mSKIPPED[0m apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
[gw2] [32mPASSED[0m apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
[gw1] [32mPASSED[0m apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
[gw1] [33mSKIPPED[0m apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification
apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
[gw0] [33mSKIPPED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
[gw4] [32mPASSED[0m apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw4] [33mSKIPPED[0m apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it
[gw2] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
[gw1] [32mPASSED[0m apache_beam/ml/inference/vertex_ai_inference_it_test.py::VertexAIInference::test_vertex_ai_run_llm_text_classification
[gw0] [32mPASSED[0m apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
[gw2] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw5] [32mPASSED[0m apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
[gw2] [32mPASSED[0m apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
[gw7] [32mPASSED[0m apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load
[gw6] [32mPASSED[0m apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit
apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
[gw6] [32mPASSED[0m apache_beam/ml/gcp/visionml_test_it.py::VisionMlTestIT::test_text_detection_with_language_hint
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_datatable_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_numpy_single_batch_large_model
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_pandas_single_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_multi_batch
apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw6] [33mSKIPPED[0m apache_beam/ml/inference/xgboost_inference_it_test.py::XGBoostInference::test_iris_classification_scipy_single_batch
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_very_selective_filters
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_iobase_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadAllBQTests::test_read_queries
apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[gw3] [32mPASSED[0m apache_beam/io/gcp/bigquery_read_it_test.py::ReadInteractiveRunnerTests::test_read_in_interactive_runner
[33m=============================== warnings summary ===============================[0m
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
../../build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py:28
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/hdfs/config.py>:28: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
from imp import load_source
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
apache_beam/ml/inference/vertex_ai_inference_it_test.py:68
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/ml/inference/vertex_ai_inference_it_test.py>:68: PytestUnknownMarkWarning: Unknown pytest.mark.uses_vertex_ai - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/stable/how-to/mark.html
@pytest.mark.uses_vertex_ai
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/dataframe/flight_delays.py>:47: FutureWarning: The default value of numeric_only in DataFrame.mean is deprecated. In a future version, it will default to False. In addition, specifying 'numeric_only=None' is deprecated. Select only valid columns or specify the value of numeric_only to silence this warning.
return airline_df[at_top_airports].mean()
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/pytest_postCommitIT-df-py38.xml> -
[33m========== [32m89 passed[0m, [33m[1m49 skipped[0m, [33m[1m26 warnings[0m[33m in 10147.96s (2:49:07)[0m[33m ===========[0m
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/portable/common.gradle'> line: 326
* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py38:postCommitPy38IT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.
For more on this, please refer to https://docs.gradle.org/8.3/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.
BUILD FAILED in 3h 30m 6s
215 actionable tasks: 153 executed, 58 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/7mke6wu7wesdo
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org