You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2024/01/24 06:32:33 UTC

Build failed in Jenkins: beam_PostCommit_Python311 #1152

See <https://ci-beam.apache.org/job/beam_PostCommit_Python311/1152/display/redirect>

Changes:


------------------------------------------
[...truncated 8.39 MB...]
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:289"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706073391
  nanos: 866988182
}
message: "Closing all cached grpc data channels."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/data_plane.py:820"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706073391
  nanos: 867060899
}
message: "Closing all cached gRPC state handlers."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:929"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706073391
  nanos: 868487119
}
message: "Done consuming work."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:301"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706073391
  nanos: 868598461
}
message: "Python sdk harness exiting."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker_main.py:213"
thread: "MainThread"

INFO:apache_beam.runners.portability.local_job_service:Completed job in 29.127021312713623 seconds with state DONE.
INFO:root:Completed job in 29.127021312713623 seconds with state DONE.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE

> Task :sdks:python:test-suites:portable:py311:portableWordCountSparkRunnerBatch
INFO:apache_beam.runners.worker.worker_pool_main:Listening for workers at localhost:43105
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7f7cfbc17e20> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f7cfbc17ec0> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7f7cfbc18720> ====================
INFO:apache_beam.utils.subprocess_server:Starting service with ('java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/runners/spark/3/job-server/build/libs/beam-runners-spark-3-job-server-2.54.0-SNAPSHOT.jar'> '--spark-master-url' 'local[4]' '--artifacts-dir' '/tmp/beam-tempz4km9sn9/artifactsu5__lki8' '--job-port' '48925' '--artifact-port' '0' '--expansion-port' '0')
WARNING:root:Waiting for grpc channel to be ready at localhost:48925.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:41 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
WARNING:root:Waiting for grpc channel to be ready at localhost:48925.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:38251
WARNING:root:Waiting for grpc channel to be ready at localhost:48925.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:44 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:33001
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:44 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:48925
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:44 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
WARNING:root:Waiting for grpc channel to be ready at localhost:48925.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:46 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_837c0823-694c-4b66-b835-a8b6caed337f.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:46 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_837c0823-694c-4b66-b835-a8b6caed337f.0:ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:46 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_837c0823-694c-4b66-b835-a8b6caed337f.null.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:46 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_837c0823-694c-4b66-b835-a8b6caed337f.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:47 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-0124051647-43ebced9_da1b5d6d-f3a3-4949-aa90-2153cea065bd
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:48 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-0124051647-43ebced9_da1b5d6d-f3a3-4949-aa90-2153cea065bd
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
  with Pipeline() as p:
    p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:48 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:50 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.util.log: Logging initialized @16636ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_392-8u392-ga-1~20.04-b08
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.Server: Started @16906ms
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@376563f3{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@684165f4{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3f10ea18{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6d615b14{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4f555e14{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@124d1bf3{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@36a1ef48{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3c9adebe{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@61ae2ec4{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4b23863c{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@be166b0{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@736cdc9b{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@25df9ed5{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@59e058b6{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7822bac9{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1ee9e0a{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5899758d{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@75fa0626{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@50c24536{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7b8459c5{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5987347c{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@b8d735c{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3c2d7c77{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5f68ded4{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5dbb463c{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:53 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2579cd0{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:54 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@73eab64c{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:54 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:54 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:54 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-0124051647-43ebced9_da1b5d6d-f3a3-4949-aa90-2153cea065bd on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:35453.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:57 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:57 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:33787.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:38643
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:57 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:57 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:59 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:59 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0124051647-43ebced9_da1b5d6d-f3a3-4949-aa90-2153cea065bd: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:59 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:59 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0124051647-43ebced9_da1b5d6d-f3a3-4949-aa90-2153cea065bd finished.
INFO:apache_beam.utils.subprocess_server:24/01/24 05:16:59 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@376563f3{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread run_worker_1-1:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
Exception in thread read_state:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38643 {created_time:"2024-01-24T05:16:59.954691589+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 266, in run
    for work_request in self._control_stub.Control(get_responses()):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:35453 {created_time:"2024-01-24T05:16:59.954747893+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 686, in <lambda>
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1060, in pull_responses
    target=lambda: self._read_inputs(elements_iterator),
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for response in responses:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:33787 {created_time:"2024-01-24T05:16:59.954729917+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38643 {created_time:"2024-01-24T05:16:59.954691589+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>

> Task :sdks:python:test-suites:portable:py311:postCommitPy311

> Task :sdks:python:test-suites:dataflow:py311:postCommitIT
<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/2050596099/bin/python3.11>: No module named build
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md

warning: check: missing required meta-data: url

warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) should be supplied


> Task :sdks:python:test-suites:dataflow:py311:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py311:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 1h 44m 6s
217 actionable tasks: 154 executed, 59 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/gor4ysr3y6vau

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python311 #1158

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python311/1158/display/redirect?page=changes>

Changes:

[noreply] Decommission all remaining jenkins jobs (#30113)


------------------------------------------
[...truncated 11.99 MB...]
           ^^^^^^^^^^^^
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py", line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38045 {created_time:"2024-01-25T17:15:24.115698433+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self._target(*self._args, **self._kwargs)
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/sdks/python/apache_beam/runners/worker/sdk_worker.py", line 1060, in pull_responses
    self._target(*self._args, **self._kwargs)
    self._target(*self._args, **self._kwargs)
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/sdks/python/apache_beam/runners/worker/data_plane.py", line 686, in <lambda>
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/sdks/python/apache_beam/runners/worker/sdk_worker.py", line 266, in run
    target=lambda: self._read_inputs(elements_iterator),
    for response in responses:
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/sdks/python/apache_beam/runners/worker/data_plane.py", line 669, in _read_inputs
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py", line 540, in __next__
    for work_request in self._control_stub.Control(get_responses()):
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py", line 540, in __next__
    for elements in elements_iterator:
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py", line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py", line 966, in _next
    return self._next()
    return self._next()
           ^^^^^^^^^^^^
           ^^^^^^^^^^^^
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py", line 966, in _next
  File "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py", line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38045 {created_time:"2024-01-25T17:15:24.115698433+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
    raise self
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:44521 {created_time:"2024-01-25T17:15:24.115703577+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:38391 {created_time:"2024-01-25T17:15:24.115698521+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>

> Task :sdks:python:test-suites:portable:py311:postCommitPy311

> Task :sdks:python:test-suites:dataflow:py311:postCommitIT

[gw6] PASSED apache_beam/dataframe/io_it_test.py::ReadUsingReadGbqTests::test_ReadGbq_export_with_project 
apache_beam/dataframe/io_it_test.py::ReadUsingReadGbqTests::test_ReadGbq_with_computation 
[gw4] PASSED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_new_types 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect 
[gw4] SKIPPED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_schema_autodetect 
apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema 
[gw2] PASSED apache_beam/transforms/enrichment_handlers/bigtable_it_test.py::TestBigTableEnrichment::test_enrichment_with_bigtable_with_timestamp 
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_specifying_only_table 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read 
[gw0] PASSED apache_beam/io/gcp/big_query_query_to_table_it_test.py::BigQueryQueryToTableIT::test_big_query_standard_sql 
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification 
[gw3] PASSED apache_beam/examples/wordcount_it_test.py::WordCountIT::test_wordcount_impersonation_it 
apache_beam/examples/wordcount_it_test.py::WordCountIT::test_wordcount_it 
[gw6] PASSED apache_beam/dataframe/io_it_test.py::ReadUsingReadGbqTests::test_ReadGbq_with_computation 
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution 
[gw6] SKIPPED apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_basic_execution 
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs 
[gw6] SKIPPED apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs 
apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement 
[gw6] SKIPPED apache_beam/testing/test_stream_it_test.py::TestStreamIntegrationTests::test_multiple_outputs_with_watermark_advancement 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_batch_copy_and_delete 
[gw6] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_batch_copy_and_delete 
apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy 
[gw6] PASSED apache_beam/io/gcp/gcsio_integration_test.py::GcsIOIntegrationTest::test_copy 
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only 
[gw2] PASSED apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_catalog_item 
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event 
[gw4] PASSED apache_beam/io/gcp/bigquery_write_it_test.py::BigQueryWriteIntegrationTests::test_big_query_write_without_schema 
apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_1M_python 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_table_schema_retrieve_with_direct_read 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source 
[gw5] PASSED apache_beam/io/gcp/bigquery_json_it_test.py::BigQueryJsonIT::test_streaming_inserts 
apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_copy 
[gw5] PASSED apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_copy 
apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_rename 
[gw5] PASSED apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_rename 
apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_rename_error 
[gw5] PASSED apache_beam/io/gcp/gcsfilesystem_integration_test.py::GcsFileSystemIntegrationTest::test_rename_error 
apache_beam/examples/fastavro_it_test.py::FastavroIT::test_avro_it 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads 
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts 
[gw0] PASSED apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification 
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model 
[gw3] PASSED apache_beam/examples/wordcount_it_test.py::WordCountIT::test_wordcount_it 
apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_aggregation 
[gw2] PASSED apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_create_user_event 
apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection 
[gw4] PASSED apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_1M_python 
apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python 
[gw0] PASSED apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_mnist_classification_large_model 
apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression 
[gw0] SKIPPED apache_beam/ml/inference/sklearn_inference_it_test.py::SklearnInference::test_sklearn_regression 
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it 
[gw2] PASSED apache_beam/ml/gcp/recommendations_ai_test_it.py::RecommendationAIIT::test_predict 
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification 
[gw6] PASSED apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_data_only 
apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction 
[gw4] PASSED apache_beam/io/gcp/bigquery_io_read_it_test.py::BigqueryIOReadIT::test_bigquery_read_custom_1M_python 
apache_beam/transforms/enrichment_it_test.py::TestEnrichment::test_http_enrichment 
[gw4] SKIPPED apache_beam/transforms/enrichment_it_test.py::TestEnrichment::test_http_enrichment 
apache_beam/transforms/enrichment_it_test.py::TestEnrichment::test_http_enrichment_custom_join 
[gw4] SKIPPED apache_beam/transforms/enrichment_it_test.py::TestEnrichment::test_http_enrichment_custom_join 
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it 
[gw3] PASSED apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_aggregation 
apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich 
[gw5] PASSED apache_beam/examples/fastavro_it_test.py::FastavroIT::test_avro_it 
apache_beam/examples/streaming_wordcount_debugging_it_test.py::StreamingWordcountDebuggingIT::test_streaming_wordcount_debugging_it 
[gw5] SKIPPED apache_beam/examples/streaming_wordcount_debugging_it_test.py::StreamingWordcountDebuggingIT::test_streaming_wordcount_debugging_it 
apache_beam/examples/streaming_wordcount_it_test.py::StreamingWordCountIT::test_streaming_wordcount_it 
[gw2] PASSED apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_deidentification 
apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows 
[gw4] PASSED apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it 
apache_beam/examples/cookbook/datastore_wordcount_it_test.py::DatastoreWordCountIT::test_datastore_wordcount_it 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts 
apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load 
[gw2] PASSED apache_beam/ml/gcp/cloud_dlp_it_test.py::CloudDLPIT::test_inspection 
apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py::JuliaSetTestIT::test_run_example_with_setup_file /home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python311/src/build/gradleenv/2050596099/bin/python3.11: No module named build
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md

warning: check: missing required meta-data: url

warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) should be supplied


[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_column_selection_and_row_restriction_rows 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime 
[gw0] PASSED apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it 
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it 
[gw6] PASSED apache_beam/io/gcp/pubsub_integration_test.py::PubSubIntegrationTest::test_streaming_with_attributes 
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances 
[gw4] PASSED apache_beam/examples/cookbook/datastore_wordcount_it_test.py::DatastoreWordCountIT::test_datastore_wordcount_it 
apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays 
[gw7] PASSED apache_beam/io/gcp/bigquery_test.py::BigQueryFileLoadsIntegrationTests::test_avro_file_load 
apache_beam/examples/complete/autocomplete_test.py::AutocompleteTest::test_autocomplete_it 
[gw5] PASSED apache_beam/examples/streaming_wordcount_it_test.py::StreamingWordCountIT::test_streaming_wordcount_it 
apache_beam/io/fileio_test.py::MatchIntegrationTest::test_transform_on_gcs 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_native_datetime 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query 
[gw2] PASSED apache_beam/examples/complete/juliaset/juliaset/juliaset_test_it.py::JuliaSetTestIT::test_run_example_with_setup_file 
apache_beam/io/gcp/datastore/v1new/datastore_write_it_test.py::DatastoreWriteIT::test_datastore_write_limit 
[gw6] PASSED apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_search_instances 
apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs 
[gw4] PASSED apache_beam/examples/dataframe/flight_delays_it_test.py::FlightDelaysTest::test_flight_delays 
apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax 
[gw4] SKIPPED apache_beam/ml/gcp/naturallanguageml_test_it.py::NaturalLanguageMlTestIT::test_analyzing_syntax 
apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it 
[gw4] SKIPPED apache_beam/runners/dataflow/dataflow_exercise_metrics_pipeline_test.py::ExerciseMetricsPipelineTest::test_metrics_it 
[gw7] PASSED apache_beam/examples/complete/autocomplete_test.py::AutocompleteTest::test_autocomplete_it 
apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context 
[gw7] SKIPPED apache_beam/ml/gcp/videointelligenceml_test_it.py::VideoIntelligenceMlTestIT::test_label_detection_with_video_context 
apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it 
[gw5] PASSED apache_beam/io/fileio_test.py::MatchIntegrationTest::test_transform_on_gcs 
apache_beam/io/parquetio_it_test.py::TestParquetIT::test_parquetio_it 
[gw3] PASSED apache_beam/examples/dataframe/taxiride_it_test.py::TaxirideIT::test_enrich 
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters 
[gw6] PASSED apache_beam/io/gcp/healthcare/dicomio_integration_test.py::DICOMIoIntegrationTest::test_dicom_store_instance_from_gcs 
apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification 
[gw6] SKIPPED apache_beam/ml/inference/onnx_inference_it_test.py::OnnxInference::test_onnx_run_inference_roberta_sentiment_classification 
[gw0] PASSED apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it 
apache_beam/examples/complete/game/user_score_it_test.py::UserScoreIT::test_user_score_it 
[gw7] PASSED apache_beam/transforms/external_it_test.py::ExternalTransformIT::test_job_python_from_python_it 
[gw5] PASSED apache_beam/io/parquetio_it_test.py::TestParquetIT::test_parquetio_it 
apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it 
[gw5] SKIPPED apache_beam/transforms/periodicsequence_it_test.py::PeriodicSequenceIT::test_periodicsequence_outputs_valid_watermarks_it 
[gw1] PASSED apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_query_and_filters 
apache_beam/io/gcp/bigquery_read_it_test.py::ReadUsingStorageApiTests::test_iobase_source_with_row_restriction FATAL: command execution failed
java.io.IOException: Backing channel 'apache-beam-jenkins-16' is disconnected.
	at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:215)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:285)
	at com.sun.proxy.$Proxy192.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1212)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1204)
	at hudson.Launcher$ProcStarter.join(Launcher.java:521)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:321)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:818)
	at hudson.model.Build$BuildExecution.build(Build.java:199)
	at hudson.model.Build$BuildExecution.doRun(Build.java:164)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:526)
	at hudson.model.Run.execute(Run.java:1841)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:44)
	at hudson.model.ResourceController.execute(ResourceController.java:101)
	at hudson.model.Executor.run(Executor.java:442)
Caused by: java.io.IOException: Unexpected termination of the channel
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:75)
Caused by: java.io.EOFException
	at java.base/java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2911)
	at java.base/java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3406)
	at java.base/java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:932)
	at java.base/java.io.ObjectInputStream.<init>(ObjectInputStream.java:375)
	at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49)
	at hudson.remoting.Command.readFrom(Command.java:142)
	at hudson.remoting.Command.readFrom(Command.java:128)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:35)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
FATAL: Channel "hudson.remoting.Channel@687bffe4:apache-beam-jenkins-16": Remote call on apache-beam-jenkins-16 failed. The channel is closing down or has closed down
java.io.EOFException
	at java.base/java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2911)
	at java.base/java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3406)
	at java.base/java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:932)
	at java.base/java.io.ObjectInputStream.<init>(ObjectInputStream.java:375)
	at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49)
	at hudson.remoting.Command.readFrom(Command.java:142)
	at hudson.remoting.Command.readFrom(Command.java:128)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:35)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Caused: java.io.IOException: Unexpected termination of the channel
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:75)
Caused: hudson.remoting.ChannelClosedException: Channel "hudson.remoting.Channel@687bffe4:apache-beam-jenkins-16": Remote call on apache-beam-jenkins-16 failed. The channel is closing down or has closed down
	at hudson.remoting.Channel.call(Channel.java:996)
	at hudson.Launcher$RemoteLauncher.kill(Launcher.java:1147)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:538)
	at hudson.model.Run.execute(Run.java:1841)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:44)
	at hudson.model.ResourceController.execute(ResourceController.java:101)
	at hudson.model.Executor.run(Executor.java:442)

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python311 #1157

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python311/1157/display/redirect>

Changes:


------------------------------------------
[...truncated 11.85 MB...]
  File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job
    return self._insert_load_job(
           ^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/utils/retry.py", line 298, in wrapper
    return fun(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert
    return self._RunMethod(
           ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 11:06:03 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
> [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']

      Worker ID: beamapp-jenkins-012511023-01250302-5xp1-harness-w0lz
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-25T11:06:04.892Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/GroupShardedRows/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/DropShardNumber+WriteWithMultipleDests/BigQueryBatchFileLoads/WriteGroupedRecordsToFile+WriteWithMultipleDests/BigQueryBatchFileLoads/DestinationFilesUnion/InputIdentity+WriteWithMultipleDests/BigQueryBatchFileLoads/IdentityWorkaround+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Write
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-25T11:06:05.218Z: JOB_MESSAGE_BASIC: Stopping worker pool...
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-25T11:08:16.570Z: JOB_MESSAGE_BASIC: Worker pool stopped.
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:151 Job 2024-01-25_03_02_35-4556418147885796967 is in state JOB_STATE_FAILED
ERROR    apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:777 Console URL: https://console.cloud.google.com/dataflow/jobs/<RegionId>/2024-01-25_03_02_35-4556418147885796967?project=<ProjectId>
INFO     apache_beam.io.gcp.bigquery_file_loads_test:bigquery_file_loads_test.py:1242 Deleting dataset python_bq_file_loads_1706180551ca2635 in project apache-beam-testing
=============================== warnings summary ===============================
apache_beam/io/gcp/bigquery.py:2634
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:2634: DeprecationWarning: invalid escape sequence '\#'
    """Read data from BigQuery.

apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    dataset_ref = client.dataset(unique_dataset_name, project=project)

apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    table_ref = client.dataset(dataset_id).table(table_id)

apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/pytest_postCommitIT-df-py311.xml> -
=========================== short test summary info ============================
FAILED apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_batch_copy_jobs_with_no_input_schema - AssertionError: 
Expected: (<[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), (3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), (8,), (9,)])>]>)
     but: <[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), (3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), (8,), (9,)])>]> was <DataflowPipelineResult 2024-01-25_02_55_42-1932701615984870965 DONE>
FAILED apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_multiple_destinations_transform - apache_beam.runners.dataflow.dataflow_runner.DataflowRuntimeException: Dataflow pipeline failed. State: FAILED, Error:
Traceback (most recent call last):
  File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job
    return self._insert_load_job(
           ^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/utils/retry.py", line 298, in wrapper
    return fun(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert
    return self._RunMethod(
           ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 11:06:03 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
>

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 311, in _execute
    response = task()
               ^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 386, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 650, in do_instruction
    return getattr(self, request_type)(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 688, in process_bundle
    bundle_processor.process_bundle(instruction_id))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1113, in process_bundle
    input_op_by_transform_id[element.transform_id].process_encoded(
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/bundle_processor.py", line 237, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 570, in apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 572, in apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1526, in apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 636, in apache_beam.runners.common.SimpleInvoker.invoke_process
  File "apache_beam/runners/common.py", line 1621, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "apache_beam/runners/common.py", line 1736, in apache_beam.runners.common._OutputHandler._write_value_to_tag
  File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 1351, in apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 1354, in apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 572, in apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1547, in apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job
    return self._insert_load_job(
           ^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/utils/retry.py", line 298, in wrapper
    return fun(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert
    return self._RunMethod(
           ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 11:06:03 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
> [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']
====== 2 failed, 89 passed, 52 skipped, 9 warnings in 5791.13s (1:36:31) =======

> Task :sdks:python:test-suites:dataflow:py311:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py311:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 1h 43m 23s
217 actionable tasks: 151 executed, 62 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/obq7vzcwycp6y

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python311 #1156

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python311/1156/display/redirect>

Changes:


------------------------------------------
[...truncated 5.55 MB...]
  seconds: 1706159793
  nanos: 7424831
}
message: "SDK Harness waiting for in-flight requests to complete"
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:289"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706159793
  nanos: 7495880
}
message: "Closing all cached grpc data channels."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/data_plane.py:820"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706159793
  nanos: 7571220
}
message: "Closing all cached gRPC state handlers."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:929"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706159793
  nanos: 8620977
}
message: "Done consuming work."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:301"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706159793
  nanos: 8751392
}
message: "Python sdk harness exiting."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker_main.py:213"
thread: "MainThread"

INFO:apache_beam.runners.portability.local_job_service:Completed job in 29.027992963790894 seconds with state DONE.
INFO:root:Completed job in 29.027992963790894 seconds with state DONE.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE

> Task :sdks:python:test-suites:portable:py311:portableWordCountSparkRunnerBatch
INFO:apache_beam.runners.worker.worker_pool_main:Listening for workers at localhost:40659
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7f75224c8180> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f75224c8220> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7f75224c8a40> ====================
INFO:apache_beam.utils.subprocess_server:Starting service with ('java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/runners/spark/3/job-server/build/libs/beam-runners-spark-3-job-server-2.55.0-SNAPSHOT.jar'> '--spark-master-url' 'local[4]' '--artifacts-dir' '/tmp/beam-temp0_dmhms3/artifactsgccf4etc' '--job-port' '47393' '--artifact-port' '0' '--expansion-port' '0')
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:43 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:35799
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:35963
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:47393
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:43 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_00b38cfc-2b0e-48b7-b62e-081120bcc3a1.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_00b38cfc-2b0e-48b7-b62e-081120bcc3a1.0:ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_00b38cfc-2b0e-48b7-b62e-081120bcc3a1.null.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:45 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_00b38cfc-2b0e-48b7-b62e-081120bcc3a1.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:46 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-0125051646-c36f0b1f_95fa490b-e0b7-457f-a0cf-ef944707fab0
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:46 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-0125051646-c36f0b1f_95fa490b-e0b7-457f-a0cf-ef944707fab0
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
  with Pipeline() as p:
    p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:46 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:47 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:47 INFO org.sparkproject.jetty.util.log: Logging initialized @7043ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_392-8u392-ga-1~20.04-b08
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.Server: Started @7162ms
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@3438b9e8{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@149907ca{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6e28e0e8{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1c5ca4c0{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@116d0880{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5cbcfb47{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@ce86f1d{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@29630d16{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@787a582{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@304dfacb{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5410083e{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5680884d{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@64bc2156{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1ac412ba{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6346d1da{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@30d89250{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2c2f0387{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4a33899f{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@712290b2{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@77e69ffb{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1b0db5e6{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5ecf3b4{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@30148804{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@59dc4a97{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@559e5595{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@50495143{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4b3324cc{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:48 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-0125051646-c36f0b1f_95fa490b-e0b7-457f-a0cf-ef944707fab0 on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:43843.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:50 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:50 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:45307.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:36995
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:50 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:50 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0125051646-c36f0b1f_95fa490b-e0b7-457f-a0cf-ef944707fab0: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:52 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0125051646-c36f0b1f_95fa490b-e0b7-457f-a0cf-ef944707fab0 finished.
INFO:apache_beam.utils.subprocess_server:24/01/25 05:16:52 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@3438b9e8{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread run_worker_1-1:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
Exception in thread read_state:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self._target(*self._args, **self._kwargs)
    self.run()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 266, in run
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1060, in pull_responses
    for work_request in self._control_stub.Control(get_responses()):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:36995 {created_time:"2024-01-25T05:16:52.655132139+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
    for response in responses:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
    return self._next()
Exception in thread read_grpc_client_inputs:
           ^^^^^^^^^^^^
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
    raise self
    self.run()
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:43843 {created_time:"2024-01-25T05:16:52.655188152+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
  File "/usr/lib/python3.11/threading.py", line 975, in run
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:45307 {grpc_message:"recvmsg:Connection reset by peer", grpc_status:14, created_time:"2024-01-25T05:16:52.655186617+00:00"}"
>
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 686, in <lambda>
    target=lambda: self._read_inputs(elements_iterator),
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:36995 {created_time:"2024-01-25T05:16:52.655132139+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>

> Task :sdks:python:test-suites:portable:py311:postCommitPy311

> Task :sdks:python:test-suites:dataflow:py311:postCommitIT
<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/2050596099/bin/python3.11>: No module named build
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md

warning: check: missing required meta-data: url

warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) should be supplied


> Task :sdks:python:test-suites:dataflow:py311:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py311:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 1h 45m 10s
217 actionable tasks: 151 executed, 62 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/kik662uz3wwqc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python311 #1155

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python311/1155/display/redirect>

Changes:


------------------------------------------
[...truncated 10.01 MB...]
}
message: "SDK Harness waiting for in-flight requests to complete"
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:289"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706138218
  nanos: 907072067
}
message: "Closing all cached grpc data channels."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/data_plane.py:820"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706138218
  nanos: 907138109
}
message: "Closing all cached gRPC state handlers."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:929"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706138218
  nanos: 908039093
}
message: "Done consuming work."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:301"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706138218
  nanos: 908141851
}
message: "Python sdk harness exiting."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker_main.py:213"
thread: "MainThread"

INFO:apache_beam.runners.portability.local_job_service:Completed job in 28.065586805343628 seconds with state DONE.
INFO:root:Completed job in 28.065586805343628 seconds with state DONE.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE

> Task :sdks:python:test-suites:portable:py311:portableWordCountSparkRunnerBatch
INFO:apache_beam.runners.worker.worker_pool_main:Listening for workers at localhost:44677
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7f7ab1930180> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f7ab1930220> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7f7ab1930a40> ====================
INFO:apache_beam.utils.subprocess_server:Starting service with ('java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/runners/spark/3/job-server/build/libs/beam-runners-spark-3-job-server-2.55.0-SNAPSHOT.jar'> '--spark-master-url' 'local[4]' '--artifacts-dir' '/tmp/beam-temp32vwj911/artifactsb_v784c1' '--job-port' '57805' '--artifact-port' '0' '--expansion-port' '0')
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:09 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:09 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:38881
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:09 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:45385
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:09 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:57805
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:09 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
WARNING:root:Waiting for grpc channel to be ready at localhost:57805.
WARNING:root:Waiting for grpc channel to be ready at localhost:57805.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:13 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_92d9aba6-7997-4a5a-8b7c-7ff2c03261b6.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:13 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_92d9aba6-7997-4a5a-8b7c-7ff2c03261b6.0:ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:13 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_92d9aba6-7997-4a5a-8b7c-7ff2c03261b6.null.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:13 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_92d9aba6-7997-4a5a-8b7c-7ff2c03261b6.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:13 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-0124231713-c8ae0832_e7855f4b-b8a1-42d7-a704-dbf1b02eadaf
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:13 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-0124231713-c8ae0832_e7855f4b-b8a1-42d7-a704-dbf1b02eadaf
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
  with Pipeline() as p:
    p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:13 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:14 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.util.log: Logging initialized @8294ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_392-8u392-ga-1~20.04-b08
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.Server: Started @8406ms
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@6d167349{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@61e9c80a{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@13a43ffd{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6cd9af9b{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@74f2dd2{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@51f2d8a1{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2616781a{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@182af2c2{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5a180a35{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@54afda32{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1edf069{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3c342fad{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@39254e47{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@32cfbba2{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5ac6a18a{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6777700d{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@55c1df0e{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@775e3108{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@315cf217{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3227286b{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7d55ad0b{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7e897a63{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4b1f4e07{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2e7ead60{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@69056340{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@63d0e8cd{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7728020c{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:15 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-0124231713-c8ae0832_e7855f4b-b8a1-42d7-a704-dbf1b02eadaf on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:39147.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:42469.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:35599
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:17 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0124231713-c8ae0832_e7855f4b-b8a1-42d7-a704-dbf1b02eadaf: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:18 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0124231713-c8ae0832_e7855f4b-b8a1-42d7-a704-dbf1b02eadaf finished.
INFO:apache_beam.utils.subprocess_server:24/01/24 23:17:19 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@6d167349{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread read_state:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
Exception in thread run_worker_1-1:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:35599 {grpc_message:"Socket closed", grpc_status:14, created_time:"2024-01-24T23:17:19.574057672+00:00"}"
>
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
    self.run()
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 686, in <lambda>
    target=lambda: self._read_inputs(elements_iterator),
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for elements in elements_iterator:
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 266, in run
    self._target(*self._args, **self._kwargs)
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1060, in pull_responses
    for work_request in self._control_stub.Control(get_responses()):
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    for response in responses:
    raise self
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:35599 {grpc_message:"Socket closed", grpc_status:14, created_time:"2024-01-24T23:17:19.574057672+00:00"}"
>
    return self._next()
    return self._next()
           ^^^^^^^^^^^^
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:39147 {created_time:"2024-01-24T23:17:19.574077368+00:00", grpc_status:14, grpc_message:"recvmsg:Connection reset by peer"}"
>
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:42469 {grpc_message:"Socket closed", grpc_status:14, created_time:"2024-01-24T23:17:19.574056711+00:00"}"
>

> Task :sdks:python:test-suites:portable:py311:postCommitPy311

> Task :sdks:python:test-suites:dataflow:py311:postCommitIT
<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/2050596099/bin/python3.11>: No module named build
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md

warning: check: missing required meta-data: url

warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) should be supplied


> Task :sdks:python:test-suites:dataflow:py311:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py311:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 1h 47m 39s
217 actionable tasks: 160 executed, 53 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/5lko2ojykop5c

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python311 #1154

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python311/1154/display/redirect>

Changes:


------------------------------------------
[...truncated 11.93 MB...]
  File "apache_beam/runners/common.py", line 1547, in apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job
    return self._insert_load_job(
           ^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/utils/retry.py", line 298, in wrapper
    return fun(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert
    return self._RunMethod(
           ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Wed, 24 Jan 2024 17:06:43 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
> [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']

      Worker ID: beamapp-jenkins-012417030-01240903-xf0p-harness-kkjp
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-24T17:06:44.848Z: JOB_MESSAGE_BASIC: Executing operation WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda at bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-24T17:06:44.922Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda at bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-24T17:06:45.272Z: JOB_MESSAGE_BASIC: Stopping worker pool...
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-24T17:08:53.080Z: JOB_MESSAGE_BASIC: Worker pool stopped.
INFO     apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:151 Job 2024-01-24_09_03_03-1581133351479077354 is in state JOB_STATE_FAILED
ERROR    apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:777 Console URL: https://console.cloud.google.com/dataflow/jobs/<RegionId>/2024-01-24_09_03_03-1581133351479077354?project=<ProjectId>
INFO     apache_beam.io.gcp.bigquery_file_loads_test:bigquery_file_loads_test.py:1242 Deleting dataset python_bq_file_loads_17061157797ccdf5 in project apache-beam-testing
=============================== warnings summary ===============================
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    dataset_ref = client.dataset(unique_dataset_name, project=project)

apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead.
    table_ref = client.dataset(dataset_id).table(table_id)

apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
  <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/pytest_postCommitIT-df-py311.xml> -
=========================== short test summary info ============================
FAILED apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_batch_copy_jobs_with_no_input_schema - AssertionError: 
Expected: (<[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), (3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), (8,), (9,)])>]>)
     but: <[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), (3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), (8,), (9,)])>]> was <DataflowPipelineResult 2024-01-24_08_55_40-3852995798183440526 DONE>
FAILED apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_multiple_destinations_transform - apache_beam.runners.dataflow.dataflow_runner.DataflowRuntimeException: Dataflow pipeline failed. State: FAILED, Error:
Traceback (most recent call last):
  File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job
    return self._insert_load_job(
           ^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/utils/retry.py", line 298, in wrapper
    return fun(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert
    return self._RunMethod(
           ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Wed, 24 Jan 2024 17:06:43 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
>

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 311, in _execute
    response = task()
               ^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 386, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 650, in do_instruction
    return getattr(self, request_type)(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py", line 688, in process_bundle
    bundle_processor.process_bundle(instruction_id))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1113, in process_bundle
    input_op_by_transform_id[element.transform_id].process_encoded(
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/runners/worker/bundle_processor.py", line 237, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 570, in apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 572, in apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1526, in apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 636, in apache_beam.runners.common.SimpleInvoker.invoke_process
  File "apache_beam/runners/common.py", line 1621, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "apache_beam/runners/common.py", line 1736, in apache_beam.runners.common._OutputHandler._write_value_to_tag
  File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 1351, in apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 1354, in apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 572, in apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1547, in apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job
    return self._insert_load_job(
           ^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/utils/retry.py", line 298, in wrapper
    return fun(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.11/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert
    return self._RunMethod(
           ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Wed, 24 Jan 2024 17:06:43 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
> [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']
====== 2 failed, 88 passed, 52 skipped, 8 warnings in 6012.56s (1:40:12) =======

> Task :sdks:python:test-suites:dataflow:py311:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py311:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 1h 47m 10s
217 actionable tasks: 154 executed, 59 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/ttafvx2eftiwy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python311 #1153

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python311/1153/display/redirect?page=changes>

Changes:

[noreply] Merge pull request #30059 [YAML] Require numStreams for unbounded


------------------------------------------
[...truncated 4.94 MB...]
  nanos: 612756252
}
message: "SDK Harness waiting for in-flight requests to complete"
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:289"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706094918
  nanos: 612830400
}
message: "Closing all cached grpc data channels."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/data_plane.py:820"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706094918
  nanos: 612891674
}
message: "Closing all cached gRPC state handlers."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:929"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706094918
  nanos: 614019632
}
message: "Done consuming work."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker.py:301"
thread: "MainThread"

INFO:root:severity: INFO
timestamp {
  seconds: 1706094918
  nanos: 614110469
}
message: "Python sdk harness exiting."
log_location: "/usr/local/lib/python3.11/site-packages/apache_beam/runners/worker/sdk_worker_main.py:213"
thread: "MainThread"

INFO:apache_beam.runners.portability.local_job_service:Completed job in 22.39385175704956 seconds with state DONE.
INFO:root:Completed job in 22.39385175704956 seconds with state DONE.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE

> Task :sdks:python:test-suites:portable:py311:portableWordCountSparkRunnerBatch
INFO:apache_beam.runners.worker.worker_pool_main:Listening for workers at localhost:36005
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function pack_combiners at 0x7f6d9cf83e20> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f6d9cf83ec0> ====================
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function sort_stages at 0x7f6d9cf84720> ====================
INFO:apache_beam.utils.subprocess_server:Starting service with ('java' '-jar' '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/runners/spark/3/job-server/build/libs/beam-runners-spark-3-job-server-2.54.0-SNAPSHOT.jar'> '--spark-master-url' 'local[4]' '--artifacts-dir' '/tmp/beam-templuedgrxg/artifactscazwj6r5' '--job-port' '45005' '--artifact-port' '0' '--expansion-port' '0')
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:26 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:26 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingService started on localhost:36455
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:26 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService started on localhost:38515
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:26 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on localhost:45005
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:26 INFO org.apache.beam.runners.jobsubmission.JobServerDriver: Job server now running, terminate with Ctrl+C
WARNING:root:Waiting for grpc channel to be ready at localhost:45005.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:28 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging artifacts for job_01e2071e-1712-4c86-b871-e110f7aef55c.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:28 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Resolving artifacts for job_01e2071e-1712-4c86-b871-e110f7aef55c.0:ref_Environment_default_environment_1.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:28 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting 1 artifacts for job_01e2071e-1712-4c86-b871-e110f7aef55c.null.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:29 INFO org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Artifacts fully staged for job_01e2071e-1712-4c86-b871-e110f7aef55c.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:29 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job BeamApp-jenkins-0124111529-731c2e8a_20fd3ca0-ddd8-499d-ac2e-e82162f918d0
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:29 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation BeamApp-jenkins-0124111529-731c2e8a_20fd3ca0-ddd8-499d-ac2e-e82162f918d0
INFO:apache_beam.runners.portability.portable_runner:Environment "LOOPBACK" has started a component necessary for the execution. Be sure to run the pipeline using
  with Pipeline() as p:
    p.apply(..)
This ensures that the pipeline finishes before this program exits.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:30 INFO org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a brand new Spark Context.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:30 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.util.log: Logging initialized @8026ms to org.sparkproject.jetty.util.log.Slf4jLog
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.Server: jetty-9.4.44.v20210927; built: 2021-09-27T23:02:44.612Z; git: 8da83308eeca865e495e53ef315a249d63ba9332; jvm 1.8.0_392-8u392-ga-1~20.04-b08
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.Server: Started @8129ms
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.AbstractConnector: Started ServerConnector@4d9dee78{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@21ef2596{/jobs,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1c052172{/jobs/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3597d7c5{/jobs/job,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@75e451eb{/jobs/job/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@24d29023{/stages,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@41cc908d{/stages/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3be8bef2{/stages/stage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@217a242d{/stages/stage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@69767cf0{/stages/pool,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1eb11ee0{/stages/pool/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@728132e9{/storage,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5ad7d181{/storage/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5b5d1518{/storage/rdd,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4f9261d9{/storage/rdd/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@5fab9c63{/environment,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@47357dba{/environment/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@459bfceb{/executors,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@18b6dc80{/executors/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4581826b{/executors/threadDump,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@aa6a584{/executors/threadDump/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@26f1994d{/static,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4a238ef5{/,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@67ddbf23{/api,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1fda33f0{/jobs/job/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:31 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6706473a{/stages/stage/kill,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:32 INFO org.sparkproject.jetty.server.handler.ContextHandler: Started o.s.j.s.ServletContextHandler@10fe8054{/metrics/json,null,AVAILABLE,@Spark}
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:32 INFO org.apache.beam.runners.spark.metrics.MetricsAccumulator: Instantiated metrics accumulator: MetricQueryResults()
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:32 WARN software.amazon.awssdk.regions.internal.util.EC2MetadataUtils: Unable to retrieve the requested metadata.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:32 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job BeamApp-jenkins-0124111529-731c2e8a_20fd3ca0-ddd8-499d-ac2e-e82162f918d0 on Spark master local[4]
INFO:apache_beam.runners.worker.statecache:Creating state cache with size 104857600
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure control channel for localhost:46351.
INFO:apache_beam.runners.worker.sdk_worker:Control channel established.
INFO:apache_beam.runners.worker.sdk_worker:Initializing SDKHarness with unbounded number of workers.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 1-1
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-2
INFO:apache_beam.runners.worker.sdk_worker:Creating insecure state channel for localhost:42321.
INFO:apache_beam.runners.worker.sdk_worker:State channel established.
INFO:apache_beam.runners.worker.data_plane:Creating client data channel for localhost:43119
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:34 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-3
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-4
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:34 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-5
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-6
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-9
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-8
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-7
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-10
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-11
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-12
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-13
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-14
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-15
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0124111529-731c2e8a_20fd3ca0-ddd8-499d-ac2e-e82162f918d0: Pipeline translated successfully. Computing outputs
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 1-16
INFO:apache_beam.io.filebasedsink:Starting finalize_write threads with num_shards: 4 (skipped: 0), batches: 4, num_threads: 4
INFO:apache_beam.io.filebasedsink:Renamed 4 shards in 0.01 seconds.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job BeamApp-jenkins-0124111529-731c2e8a_20fd3ca0-ddd8-499d-ac2e-e82162f918d0 finished.
INFO:apache_beam.utils.subprocess_server:24/01/24 11:15:35 INFO org.sparkproject.jetty.server.AbstractConnector: Stopped Spark@4d9dee78{HTTP/1.1, (http/1.1)}{127.0.0.1:4040}
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
Exception in thread read_state:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
Exception in thread run_worker_1-1:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self.run()
  File "/usr/lib/python3.11/threading.py", line 975, in run
    self._target(*self._args, **self._kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 1060, in pull_responses
    for response in responses:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    self._target(*self._args, **self._kwargs)
    return self._next()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker.py",> line 266, in run
ERROR:apache_beam.runners.worker.data_plane:Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:43119 {grpc_message:"recvmsg:Connection reset by peer", grpc_status:14, created_time:"2024-01-24T11:15:36.224866387+00:00"}"
>
           ^^^^^^^^^^^^
    for work_request in self._control_stub.Control(get_responses()):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
Exception in thread read_grpc_client_inputs:
Traceback (most recent call last):
  File "/usr/lib/python3.11/threading.py", line 1038, in _bootstrap_inner
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    self.run()
    raise self
  File "/usr/lib/python3.11/threading.py", line 975, in run
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:46351 {grpc_message:"recvmsg:Connection reset by peer", grpc_status:14, created_time:"2024-01-24T11:15:36.22487283+00:00"}"
>
    raise self
    self._target(*self._args, **self._kwargs)
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "Socket closed"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:42321 {created_time:"2024-01-24T11:15:36.224869811+00:00", grpc_status:14, grpc_message:"Socket closed"}"
>
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 686, in <lambda>
    target=lambda: self._read_inputs(elements_iterator),
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/apache_beam/runners/worker/data_plane.py",> line 669, in _read_inputs
    for elements in elements_iterator:
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 540, in __next__
    return self._next()
           ^^^^^^^^^^^^
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/-1720702906/lib/python3.11/site-packages/grpc/_channel.py",> line 966, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "recvmsg:Connection reset by peer"
	debug_error_string = "UNKNOWN:Error received from peer ipv6:%5B::1%5D:43119 {grpc_message:"recvmsg:Connection reset by peer", grpc_status:14, created_time:"2024-01-24T11:15:36.224866387+00:00"}"
>

> Task :sdks:python:test-suites:portable:py311:postCommitPy311

> Task :sdks:python:test-suites:dataflow:py311:postCommitIT
<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/build/gradleenv/2050596099/bin/python3.11>: No module named build
warning: sdist: standard file not found: should have one of README, README.rst, README.txt, README.md

warning: check: missing required meta-data: url

warning: check: missing meta-data: either (author and author_email) or (maintainer and maintainer_email) should be supplied


> Task :sdks:python:test-suites:dataflow:py311:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python311/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py311:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

For more on this, please refer to https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation.

BUILD FAILED in 1h 44m 47s
217 actionable tasks: 154 executed, 59 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/u775je4cooejg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org