You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/05/09 06:06:57 UTC

Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3134

See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3134/display/redirect?page=changes>

Changes:

[zyichi] [BEAM-9940] Set timer family spec for TimerDeclarations in dataflow


------------------------------------------
[...truncated 574.40 KB...]
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f6aa26d3e60> ====================
ERROR:grpc._channel:Exception iterating requests!
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 195, in consume_request_iterator
    request = next(request_iterator)
  File "apache_beam/runners/portability/artifact_service.py", line 317, in __next__
    raise self._queue.get()
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNIMPLEMENTED
	details = "Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService"
	debug_error_string = "{"created":"@1589004411.541735884","description":"Error received from peer ipv4:127.0.0.1:35745","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService","grpc_status":12}"
>
20/05/09 06:06:51 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589004411.09_3541fb94-94f0-4d28-9205-c62bb962332c
20/05/09 06:06:51 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589004411.09_3541fb94-94f0-4d28-9205-c62bb962332c
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/09 06:06:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/09 06:06:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/09 06:06:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589004411.09_3541fb94-94f0-4d28-9205-c62bb962332c on Spark master local
20/05/09 06:06:51 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/09 06:06:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589004411.09_3541fb94-94f0-4d28-9205-c62bb962332c: Pipeline translated successfully. Computing outputs
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/09 06:06:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/09 06:06:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:41883
20/05/09 06:06:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/09 06:06:52 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/09 06:06:52 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589004411.09_3541fb94-94f0-4d28-9205-c62bb962332c', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/09 06:06:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589004411.09', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:53043', 'job_port': u'0'}
20/05/09 06:06:52 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/09 06:06:52 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:34377.
20/05/09 06:06:52 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/09 06:06:52 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/09 06:06:52 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:43119.
20/05/09 06:06:52 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/09 06:06:52 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/09 06:06:52 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589004412.412176431","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589004412.412173165","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589004412.412171980","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589004412.412166032","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589004412.412150726","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589004412.412176431","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589004412.412173165","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589004412.412171980","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589004412.412166032","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589004412.412150726","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/09 06:06:52 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:40741
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/09 06:06:52 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/09 06:06:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/09 06:06:52 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589004411.09_3541fb94-94f0-4d28-9205-c62bb962332c finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f6aa26d3e60> ====================
ERROR:grpc._channel:Exception iterating requests!
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 195, in consume_request_iterator
    request = next(request_iterator)
  File "apache_beam/runners/portability/artifact_service.py", line 317, in __next__
    raise self._queue.get()
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNIMPLEMENTED
	details = "Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService"
	debug_error_string = "{"created":"@1589004413.506945135","description":"Error received from peer ipv4:127.0.0.1:35745","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService","grpc_status":12}"
>
20/05/09 06:06:53 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589004412.92_a0014ac6-339f-4d35-9621-7c5f03e89643
20/05/09 06:06:53 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589004412.92_a0014ac6-339f-4d35-9621-7c5f03e89643
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/09 06:06:53 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/09 06:06:53 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/09 06:06:53 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589004412.92_a0014ac6-339f-4d35-9621-7c5f03e89643 on Spark master local
20/05/09 06:06:53 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/09 06:06:53 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589004412.92_a0014ac6-339f-4d35-9621-7c5f03e89643: Pipeline translated successfully. Computing outputs
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/09 06:06:54 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/09 06:06:54 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:36633
20/05/09 06:06:54 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/09 06:06:54 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/09 06:06:54 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589004412.92_a0014ac6-339f-4d35-9621-7c5f03e89643', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/09 06:06:54 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589004412.92', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:53043', 'job_port': u'0'}
20/05/09 06:06:54 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/09 06:06:54 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:35445.
20/05/09 06:06:54 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/09 06:06:54 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/09 06:06:54 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:40009.
20/05/09 06:06:54 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/09 06:06:54 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/09 06:06:54 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589004414.368134729","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589004414.368131241","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589004414.368130034","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589004414.368123642","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589004414.368109025","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589004414.368134729","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589004414.368131241","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589004414.368130034","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589004414.368123642","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589004414.368109025","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/09 06:06:54 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:35505
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/09 06:06:54 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/09 06:06:54 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/09 06:06:54 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589004412.92_a0014ac6-339f-4d35-9621-7c5f03e89643 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589004366.64_e323db76-77ff-4b83-b2b1-1f54ecfee6df failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 74.038s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 28s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/en6zjsvefmcn2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Python_VR_Spark #3145

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3145/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3144

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3144/display/redirect?page=changes>

Changes:

[kamil.wasilewski] Fix InfluxDB credentials binding


------------------------------------------
[...truncated 564.10 KB...]
	at org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
	... 3 more
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 3: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 245, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 302, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 471, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 506, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7fc7c7143e60> ====================
20/05/11 18:10:24 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589220624.1_3d6b4f11-bec2-455d-aa4f-de4e35ff3c91
20/05/11 18:10:24 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589220624.1_3d6b4f11-bec2-455d-aa4f-de4e35ff3c91
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 18:10:24 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 18:10:24 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 18:10:24 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589220624.1_3d6b4f11-bec2-455d-aa4f-de4e35ff3c91 on Spark master local
20/05/11 18:10:24 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 18:10:24 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589220624.1_3d6b4f11-bec2-455d-aa4f-de4e35ff3c91: Pipeline translated successfully. Computing outputs
20/05/11 18:10:24 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 18:10:24 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 18:10:24 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 18:10:25 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 18:10:25 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 18:10:25 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:41811
20/05/11 18:10:25 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 18:10:25 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 18:10:25 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589220624.1_3d6b4f11-bec2-455d-aa4f-de4e35ff3c91', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/11 18:10:25 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589220624.1', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:56403', 'job_port': u'0'}
20/05/11 18:10:25 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 18:10:25 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:34989.
20/05/11 18:10:25 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 18:10:25 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/11 18:10:25 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 18:10:25 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/11 18:10:25 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:36739.
20/05/11 18:10:25 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 18:10:25 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 18:10:25 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589220625.555436301","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589220625.555432920","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589220625.555431732","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589220625.555425596","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589220625.555409843","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589220625.555436301","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589220625.555432920","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589220625.555431732","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589220625.555425596","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589220625.555409843","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 18:10:25 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:36095
20/05/11 18:10:25 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 18:10:25 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/11 18:10:25 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/11 18:10:25 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/11 18:10:26 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589220624.1_3d6b4f11-bec2-455d-aa4f-de4e35ff3c91 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7fc7c7143e60> ====================
20/05/11 18:10:26 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589220626.06_c4384b9a-a1a5-43b5-b789-a2c8d73ea07c
20/05/11 18:10:26 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589220626.06_c4384b9a-a1a5-43b5-b789-a2c8d73ea07c
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 18:10:26 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 18:10:26 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 18:10:26 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589220626.06_c4384b9a-a1a5-43b5-b789-a2c8d73ea07c on Spark master local
20/05/11 18:10:26 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 18:10:26 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589220626.06_c4384b9a-a1a5-43b5-b789-a2c8d73ea07c: Pipeline translated successfully. Computing outputs
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 18:10:27 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 18:10:27 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 18:10:27 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:42071
20/05/11 18:10:27 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 18:10:27 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 18:10:27 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589220626.06_c4384b9a-a1a5-43b5-b789-a2c8d73ea07c', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/11 18:10:27 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589220626.06', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:56403', 'job_port': u'0'}
20/05/11 18:10:27 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 18:10:27 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:36977.
20/05/11 18:10:27 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/11 18:10:27 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/11 18:10:27 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:44561.
20/05/11 18:10:27 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 18:10:27 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 18:10:27 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589220627.617263281","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589220627.617257824","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589220627.617255875","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589220627.617247092","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589220627.617219018","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589220627.617263281","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589220627.617257824","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589220627.617255875","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589220627.617247092","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589220627.617219018","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 18:10:27 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:39301
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/11 18:10:27 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/11 18:10:27 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589220626.06_c4384b9a-a1a5-43b5-b789-a2c8d73ea07c finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589220575.96_10748b86-05a7-4ceb-96a7-390cdb83fce3 failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 80.535s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 4m 9s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/4nhjfqjl4od7w

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3143

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3143/display/redirect?page=changes>

Changes:

[sebastian.graca] [BEAM-9439] Return split instead of total backlog size


------------------------------------------
[...truncated 560.96 KB...]
	at org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
	... 3 more
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 3: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 245, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 302, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 471, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 506, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f0ef728ee60> ====================
20/05/11 12:09:01 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589198941.25_b60dcb9b-a176-4154-a558-4ef09975b486
20/05/11 12:09:01 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589198941.25_b60dcb9b-a176-4154-a558-4ef09975b486
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 12:09:01 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 12:09:01 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 12:09:01 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589198941.25_b60dcb9b-a176-4154-a558-4ef09975b486 on Spark master local
20/05/11 12:09:01 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 12:09:01 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589198941.25_b60dcb9b-a176-4154-a558-4ef09975b486: Pipeline translated successfully. Computing outputs
20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 12:09:02 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 12:09:02 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:35089
20/05/11 12:09:02 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 12:09:02 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 12:09:02 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589198941.25_b60dcb9b-a176-4154-a558-4ef09975b486', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/11 12:09:02 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589198941.25', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:33163', 'job_port': u'0'}
20/05/11 12:09:02 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 12:09:02 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:38255.
20/05/11 12:09:02 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/11 12:09:02 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/11 12:09:02 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:46587.
20/05/11 12:09:02 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 12:09:02 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 12:09:02 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589198942.604935276","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589198942.604931879","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589198942.604930608","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589198942.604923977","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589198942.604907652","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589198942.604935276","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589198942.604931879","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589198942.604930608","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589198942.604923977","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589198942.604907652","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 12:09:02 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:42697
20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 12:09:02 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/11 12:09:02 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/11 12:09:03 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/11 12:09:03 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589198941.25_b60dcb9b-a176-4154-a558-4ef09975b486 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f0ef728ee60> ====================
20/05/11 12:09:03 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589198943.11_645e176b-9747-4f35-b245-ac879f92be67
20/05/11 12:09:03 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589198943.11_645e176b-9747-4f35-b245-ac879f92be67
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 12:09:03 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 12:09:03 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 12:09:03 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589198943.11_645e176b-9747-4f35-b245-ac879f92be67 on Spark master local
20/05/11 12:09:03 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 12:09:03 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589198943.11_645e176b-9747-4f35-b245-ac879f92be67: Pipeline translated successfully. Computing outputs
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 12:09:04 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 12:09:04 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 12:09:04 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:33877
20/05/11 12:09:04 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 12:09:04 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 12:09:04 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589198943.11_645e176b-9747-4f35-b245-ac879f92be67', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/11 12:09:04 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589198943.11', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:33163', 'job_port': u'0'}
20/05/11 12:09:04 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 12:09:04 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:38155.
20/05/11 12:09:04 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/11 12:09:04 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/11 12:09:04 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:34805.
20/05/11 12:09:04 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 12:09:04 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 12:09:04 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589198944.656603526","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589198944.656598700","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589198944.656597073","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589198944.656588465","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589198944.656566382","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589198944.656603526","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589198944.656598700","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589198944.656597073","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589198944.656588465","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589198944.656566382","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 12:09:04 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:33865
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/11 12:09:04 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/11 12:09:04 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589198943.11_645e176b-9747-4f35-b245-ac879f92be67 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589198894.96_56b42bb1-ba9e-4440-9ac7-c44afe025ae0 failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 77.047s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 39s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/7ywkfupzbbvry

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3142

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3142/display/redirect>

Changes:


------------------------------------------
[...truncated 569.78 KB...]
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sss20/05/11 06:06:29 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 06:06:29 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 06:06:29 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f988a6b5e60> ====================
20/05/11 06:06:29 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589177189.0_bdfa9130-c52e-4f16-bfff-25f986aecdda
20/05/11 06:06:29 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589177189.0_bdfa9130-c52e-4f16-bfff-25f986aecdda
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 06:06:29 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 06:06:29 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 06:06:29 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589177189.0_bdfa9130-c52e-4f16-bfff-25f986aecdda on Spark master local
20/05/11 06:06:29 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 06:06:29 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589177189.0_bdfa9130-c52e-4f16-bfff-25f986aecdda: Pipeline translated successfully. Computing outputs
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 06:06:30 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 06:06:30 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:37237
20/05/11 06:06:30 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 06:06:30 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 06:06:30 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589177189.0_bdfa9130-c52e-4f16-bfff-25f986aecdda', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/11 06:06:30 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589177189.0', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:33189', 'job_port': u'0'}
20/05/11 06:06:30 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 06:06:30 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:40883.
20/05/11 06:06:30 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/11 06:06:30 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/11 06:06:30 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:43537.
20/05/11 06:06:30 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 06:06:30 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 06:06:30 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589177190.308461719","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589177190.308458513","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589177190.308457187","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589177190.308451504","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589177190.308436214","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589177190.308461719","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589177190.308458513","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589177190.308457187","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589177190.308451504","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589177190.308436214","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 06:06:30 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:45565
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 06:06:30 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 06:06:30 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/11 06:06:30 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589177189.0_bdfa9130-c52e-4f16-bfff-25f986aecdda finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f988a6b5e60> ====================
20/05/11 06:06:31 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589177190.8_a73b092f-e523-4021-9391-a7d2df0dbb4a
20/05/11 06:06:31 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589177190.8_a73b092f-e523-4021-9391-a7d2df0dbb4a
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 06:06:31 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 06:06:31 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 06:06:31 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589177190.8_a73b092f-e523-4021-9391-a7d2df0dbb4a on Spark master local
20/05/11 06:06:31 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 06:06:31 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589177190.8_a73b092f-e523-4021-9391-a7d2df0dbb4a: Pipeline translated successfully. Computing outputs
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 06:06:32 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 06:06:32 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:46457
20/05/11 06:06:32 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 06:06:32 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 06:06:32 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589177190.8_a73b092f-e523-4021-9391-a7d2df0dbb4a', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/11 06:06:32 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589177190.8', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:33189', 'job_port': u'0'}
20/05/11 06:06:32 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 06:06:32 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:44833.
20/05/11 06:06:32 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/11 06:06:32 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/11 06:06:32 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:40177.
20/05/11 06:06:32 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 06:06:32 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 06:06:32 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589177192.292086046","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589177192.292081385","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589177192.292079810","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589177192.292071377","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589177192.292048344","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589177192.292086046","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589177192.292081385","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589177192.292079810","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589177192.292071377","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589177192.292048344","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 06:06:32 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:45527
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 06:06:32 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 06:06:32 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/11 06:06:32 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589177190.8_a73b092f-e523-4021-9391-a7d2df0dbb4a finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589177144.96_b9042930-a338-446e-963b-7120f6936490 failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 73.379s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 28s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/robyl3zztqjsw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3141

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3141/display/redirect>

Changes:


------------------------------------------
[...truncated 566.20 KB...]
	at org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
	... 3 more
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 3: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 245, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 302, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 471, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 506, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f100407de60> ====================
20/05/11 00:07:10 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589155630.0_6ffc2e4b-77a8-4b7a-bdcb-ae9663577cbd
20/05/11 00:07:10 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589155630.0_6ffc2e4b-77a8-4b7a-bdcb-ae9663577cbd
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 00:07:10 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 00:07:10 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 00:07:10 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589155630.0_6ffc2e4b-77a8-4b7a-bdcb-ae9663577cbd on Spark master local
20/05/11 00:07:10 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 00:07:10 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589155630.0_6ffc2e4b-77a8-4b7a-bdcb-ae9663577cbd: Pipeline translated successfully. Computing outputs
20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 00:07:11 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 00:07:11 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:35603
20/05/11 00:07:11 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 00:07:11 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 00:07:11 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589155630.0_6ffc2e4b-77a8-4b7a-bdcb-ae9663577cbd', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/11 00:07:11 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589155630.0', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:59757', 'job_port': u'0'}
20/05/11 00:07:11 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 00:07:11 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:38747.
20/05/11 00:07:11 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/11 00:07:11 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/11 00:07:11 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:45935.
20/05/11 00:07:11 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 00:07:11 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 00:07:11 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:44463
20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 00:07:11 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589155631.388560146","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589155631.388556731","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589155631.388555207","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589155631.388549340","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589155631.388534322","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589155631.388560146","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589155631.388556731","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589155631.388555207","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589155631.388549340","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589155631.388534322","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 00:07:11 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/11 00:07:11 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/11 00:07:11 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589155630.0_6ffc2e4b-77a8-4b7a-bdcb-ae9663577cbd finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f100407de60> ====================
20/05/11 00:07:12 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589155631.86_b0fc39c9-b02e-477f-8bf1-385481e3e72a
20/05/11 00:07:12 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589155631.86_b0fc39c9-b02e-477f-8bf1-385481e3e72a
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/11 00:07:12 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/11 00:07:12 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/11 00:07:12 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589155631.86_b0fc39c9-b02e-477f-8bf1-385481e3e72a on Spark master local
20/05/11 00:07:12 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/11 00:07:12 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589155631.86_b0fc39c9-b02e-477f-8bf1-385481e3e72a: Pipeline translated successfully. Computing outputs
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/11 00:07:13 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/11 00:07:13 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/11 00:07:13 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:36779
20/05/11 00:07:13 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/11 00:07:13 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/11 00:07:13 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589155631.86_b0fc39c9-b02e-477f-8bf1-385481e3e72a', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/11 00:07:13 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589155631.86', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:59757', 'job_port': u'0'}
20/05/11 00:07:13 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/11 00:07:13 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:40583.
20/05/11 00:07:13 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/11 00:07:13 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/11 00:07:13 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:36955.
20/05/11 00:07:13 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/11 00:07:13 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/11 00:07:13 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:38329
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/11 00:07:13 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589155633.372725920","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589155633.372720962","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589155633.372719141","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589155633.372710541","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589155633.372684706","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589155633.372725920","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589155633.372720962","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589155633.372719141","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589155633.372710541","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589155633.372684706","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/11 00:07:13 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/11 00:07:13 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589155631.86_b0fc39c9-b02e-477f-8bf1-385481e3e72a finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589155584.18_f3fb5086-f3f0-48bb-b660-2603189d8238 failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 77.939s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 31s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/rn4twkz27ocvc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3140

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3140/display/redirect>

Changes:


------------------------------------------
[...truncated 578.62 KB...]
	at org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
	... 3 more
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 3: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 245, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 302, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 471, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 506, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f1ced4f9e60> ====================
20/05/10 18:08:44 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589134123.65_8c8c6bfa-b116-4251-875f-ac5452a4748c
20/05/10 18:08:44 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589134123.65_8c8c6bfa-b116-4251-875f-ac5452a4748c
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 18:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 18:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 18:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589134123.65_8c8c6bfa-b116-4251-875f-ac5452a4748c on Spark master local
20/05/10 18:08:44 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 18:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589134123.65_8c8c6bfa-b116-4251-875f-ac5452a4748c: Pipeline translated successfully. Computing outputs
20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 18:08:45 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 18:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 18:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:37365
20/05/10 18:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 18:08:45 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 18:08:45 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589134123.65_8c8c6bfa-b116-4251-875f-ac5452a4748c', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/10 18:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589134123.65', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:46421', 'job_port': u'0'}
20/05/10 18:08:45 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 18:08:45 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:45553.
20/05/10 18:08:45 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/10 18:08:45 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/10 18:08:45 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:33959.
20/05/10 18:08:45 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 18:08:45 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 18:08:45 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589134125.504816669","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589134125.504812504","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589134125.504811253","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589134125.504803155","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589134125.504778435","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589134125.504816669","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589134125.504812504","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589134125.504811253","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589134125.504803155","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589134125.504778435","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 18:08:45 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:41499
20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/10 18:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/10 18:08:46 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/10 18:08:46 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589134123.65_8c8c6bfa-b116-4251-875f-ac5452a4748c finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f1ced4f9e60> ====================
20/05/10 18:08:46 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589134126.13_40850a38-5efe-449b-b3ae-b8b53ab3dc57
20/05/10 18:08:46 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589134126.13_40850a38-5efe-449b-b3ae-b8b53ab3dc57
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 18:08:47 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 18:08:47 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 18:08:47 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589134126.13_40850a38-5efe-449b-b3ae-b8b53ab3dc57 on Spark master local
20/05/10 18:08:47 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 18:08:47 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589134126.13_40850a38-5efe-449b-b3ae-b8b53ab3dc57: Pipeline translated successfully. Computing outputs
20/05/10 18:08:47 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 18:08:47 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 18:08:47 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/10 18:08:47 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 18:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 18:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:41479
20/05/10 18:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 18:08:47 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 18:08:47 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589134126.13_40850a38-5efe-449b-b3ae-b8b53ab3dc57', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/10 18:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589134126.13', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:46421', 'job_port': u'0'}
20/05/10 18:08:47 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 18:08:47 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:45619.
20/05/10 18:08:47 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 18:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/10 18:08:47 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 18:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/10 18:08:47 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:37843.
20/05/10 18:08:47 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 18:08:47 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 18:08:47 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:35765
20/05/10 18:08:48 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 18:08:48 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589134127.995344720","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589134127.995341062","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589134127.995339982","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589134127.995333342","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589134127.995316699","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589134127.995344720","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589134127.995341062","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589134127.995339982","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589134127.995333342","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589134127.995316699","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 18:08:48 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/10 18:08:48 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/10 18:08:48 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/10 18:08:48 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/10 18:08:48 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589134126.13_40850a38-5efe-449b-b3ae-b8b53ab3dc57 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589134066.68_7dbf1d65-4dd6-4770-9cfd-ca8c309c0dbe failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 96.208s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 4m 31s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/hxdzdn7z2wxcc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3139

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3139/display/redirect?page=changes>

Changes:

[mxm] [BEAM-9930] Beam Summit Digital 2020 announcement on blog

[mxm] Correct URL in blog post

[matthias.baetens] Minor (markup) changes

[mxm] [BEAM-9930] Correct authors of beam summit digital blog post


------------------------------------------
[...truncated 566.45 KB...]
	at org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
	... 3 more
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 3: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 245, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 302, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 471, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 506, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f6cc6ac6e60> ====================
20/05/10 12:07:48 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589112468.42_ac3fd810-8289-4bf6-800c-cd0f796ac1a0
20/05/10 12:07:48 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589112468.42_ac3fd810-8289-4bf6-800c-cd0f796ac1a0
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 12:07:49 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 12:07:49 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 12:07:49 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/10 12:07:49 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 12:07:49 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 12:07:49 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589112468.42_ac3fd810-8289-4bf6-800c-cd0f796ac1a0 on Spark master local
20/05/10 12:07:49 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 12:07:49 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589112468.42_ac3fd810-8289-4bf6-800c-cd0f796ac1a0: Pipeline translated successfully. Computing outputs
20/05/10 12:07:49 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 12:07:49 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 12:07:49 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:44351
20/05/10 12:07:49 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 12:07:49 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 12:07:49 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589112468.42_ac3fd810-8289-4bf6-800c-cd0f796ac1a0', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/10 12:07:49 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589112468.42', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:54729', 'job_port': u'0'}
20/05/10 12:07:49 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 12:07:49 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:39553.
20/05/10 12:07:49 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 12:07:49 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/10 12:07:49 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 12:07:49 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/10 12:07:49 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:37269.
20/05/10 12:07:49 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 12:07:49 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 12:07:49 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589112469.902562521","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589112469.902559238","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589112469.902558004","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589112469.902551502","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589112469.902536875","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589112469.902562521","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589112469.902559238","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589112469.902558004","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589112469.902551502","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589112469.902536875","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 12:07:49 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:46241
20/05/10 12:07:49 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 12:07:50 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/10 12:07:50 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/10 12:07:50 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/10 12:07:50 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589112468.42_ac3fd810-8289-4bf6-800c-cd0f796ac1a0 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f6cc6ac6e60> ====================
20/05/10 12:07:51 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589112470.44_64e7c370-8c2b-4703-bce8-691433d8c33b
20/05/10 12:07:51 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589112470.44_64e7c370-8c2b-4703-bce8-691433d8c33b
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 12:07:51 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 12:07:51 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 12:07:51 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/10 12:07:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 12:07:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 12:07:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589112470.44_64e7c370-8c2b-4703-bce8-691433d8c33b on Spark master local
20/05/10 12:07:51 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 12:07:51 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589112470.44_64e7c370-8c2b-4703-bce8-691433d8c33b: Pipeline translated successfully. Computing outputs
20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 12:07:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 12:07:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:43859
20/05/10 12:07:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 12:07:52 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 12:07:52 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589112470.44_64e7c370-8c2b-4703-bce8-691433d8c33b', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/10 12:07:52 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589112470.44', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:54729', 'job_port': u'0'}
20/05/10 12:07:52 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 12:07:52 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:40551.
20/05/10 12:07:52 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/10 12:07:52 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/10 12:07:52 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:33167.
20/05/10 12:07:52 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 12:07:52 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 12:07:52 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:42519
20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 12:07:52 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589112472.090841933","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589112472.090836913","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589112472.090834868","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589112472.090825268","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589112472.090791536","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589112472.090841933","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589112472.090836913","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589112472.090834868","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589112472.090825268","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589112472.090791536","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/10 12:07:52 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/10 12:07:52 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589112470.44_64e7c370-8c2b-4703-bce8-691433d8c33b finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589112418.78_de967715-018f-4d94-8386-871f013fea7f failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 83.292s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 46s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/ilrtansz7svem

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3138

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3138/display/redirect>

Changes:


------------------------------------------
[...truncated 578.32 KB...]
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f2cae347e60> ====================
ERROR:grpc._channel:Exception iterating requests!
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 195, in consume_request_iterator
    request = next(request_iterator)
  File "apache_beam/runners/portability/artifact_service.py", line 317, in __next__
    raise self._queue.get()
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNIMPLEMENTED
	details = "Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService"
	debug_error_string = "{"created":"@1589090899.174055269","description":"Error received from peer ipv4:127.0.0.1:38209","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService","grpc_status":12}"
>
20/05/10 06:08:19 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589090898.69_1fd62df6-0216-421d-a5b5-a954ed777f82
20/05/10 06:08:19 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589090898.69_1fd62df6-0216-421d-a5b5-a954ed777f82
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 06:08:19 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 06:08:19 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 06:08:19 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/10 06:08:19 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 06:08:19 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 06:08:19 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589090898.69_1fd62df6-0216-421d-a5b5-a954ed777f82 on Spark master local
20/05/10 06:08:19 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 06:08:19 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589090898.69_1fd62df6-0216-421d-a5b5-a954ed777f82: Pipeline translated successfully. Computing outputs
20/05/10 06:08:20 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 06:08:20 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 06:08:20 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:39635
20/05/10 06:08:20 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 06:08:20 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 06:08:20 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589090898.69_1fd62df6-0216-421d-a5b5-a954ed777f82', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/10 06:08:20 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589090898.69', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:59477', 'job_port': u'0'}
20/05/10 06:08:20 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 06:08:20 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:36083.
20/05/10 06:08:20 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 06:08:20 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 06:08:20 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/10 06:08:20 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/10 06:08:20 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:45509.
20/05/10 06:08:20 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 06:08:20 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 06:08:20 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:39711
20/05/10 06:08:20 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 06:08:20 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589090900.139954084","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589090900.139950304","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589090900.139948936","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589090900.139942975","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589090900.139927467","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589090900.139954084","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589090900.139950304","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589090900.139948936","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589090900.139942975","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589090900.139927467","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 06:08:20 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/10 06:08:20 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/10 06:08:20 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/10 06:08:20 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589090898.69_1fd62df6-0216-421d-a5b5-a954ed777f82 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
20/05/10 06:08:21 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 06:08:21 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 06:08:21 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f2cae347e60> ====================
ERROR:grpc._channel:Exception iterating requests!
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 195, in consume_request_iterator
    request = next(request_iterator)
  File "apache_beam/runners/portability/artifact_service.py", line 317, in __next__
    raise self._queue.get()
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNIMPLEMENTED
	details = "Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService"
	debug_error_string = "{"created":"@1589090901.451426108","description":"Error received from peer ipv4:127.0.0.1:38209","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService","grpc_status":12}"
>
20/05/10 06:08:21 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589090900.61_2d7e20d0-4329-4609-b6dc-9fe26630c682
20/05/10 06:08:21 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589090900.61_2d7e20d0-4329-4609-b6dc-9fe26630c682
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 06:08:21 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 06:08:21 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 06:08:21 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589090900.61_2d7e20d0-4329-4609-b6dc-9fe26630c682 on Spark master local
20/05/10 06:08:21 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 06:08:21 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589090900.61_2d7e20d0-4329-4609-b6dc-9fe26630c682: Pipeline translated successfully. Computing outputs
20/05/10 06:08:22 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 06:08:22 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 06:08:22 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:37901
20/05/10 06:08:22 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 06:08:22 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 06:08:22 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589090900.61_2d7e20d0-4329-4609-b6dc-9fe26630c682', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/10 06:08:22 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589090900.61', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:59477', 'job_port': u'0'}
20/05/10 06:08:22 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 06:08:22 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:40089.
20/05/10 06:08:22 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 06:08:22 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/10 06:08:22 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 06:08:22 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/10 06:08:22 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:44983.
20/05/10 06:08:22 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 06:08:22 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 06:08:22 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589090902.595560848","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589090902.595555124","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589090902.595553354","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589090902.595542587","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589090902.595474962","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589090902.595560848","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589090902.595555124","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589090902.595553354","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589090902.595542587","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589090902.595474962","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 06:08:22 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:44483
20/05/10 06:08:22 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 06:08:22 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/10 06:08:22 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/10 06:08:22 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/10 06:08:23 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/10 06:08:23 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589090900.61_2d7e20d0-4329-4609-b6dc-9fe26630c682 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589090846.84_5721944c-9529-49fa-8daa-b8b3bf2ac8c2 failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 84.755s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 56s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/pjryt3mhghkoy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3137

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3137/display/redirect>

Changes:


------------------------------------------
[...truncated 570.86 KB...]
	at org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
	... 3 more
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 3: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 245, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 302, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 471, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 506, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7faf65a49e60> ====================
20/05/10 00:08:43 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589069323.39_bace7a35-1d58-4497-94ad-bfefc0b1de82
20/05/10 00:08:43 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589069323.39_bace7a35-1d58-4497-94ad-bfefc0b1de82
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 00:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 00:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 00:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589069323.39_bace7a35-1d58-4497-94ad-bfefc0b1de82 on Spark master local
20/05/10 00:08:44 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 00:08:44 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589069323.39_bace7a35-1d58-4497-94ad-bfefc0b1de82: Pipeline translated successfully. Computing outputs
20/05/10 00:08:44 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 00:08:44 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 00:08:44 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/10 00:08:45 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 00:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 00:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:35615
20/05/10 00:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 00:08:45 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 00:08:45 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589069323.39_bace7a35-1d58-4497-94ad-bfefc0b1de82', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/10 00:08:45 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589069323.39', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:46695', 'job_port': u'0'}
20/05/10 00:08:45 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 00:08:45 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:35867.
20/05/10 00:08:45 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 00:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/10 00:08:45 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 00:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/10 00:08:45 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:44225.
20/05/10 00:08:45 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 00:08:45 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 00:08:45 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589069325.073051730","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589069325.073048237","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589069325.073047045","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589069325.073040650","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589069325.073025444","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589069325.073051730","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589069325.073048237","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589069325.073047045","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589069325.073040650","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589069325.073025444","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 00:08:45 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:40367
20/05/10 00:08:45 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 00:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/10 00:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/10 00:08:45 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/10 00:08:45 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589069323.39_bace7a35-1d58-4497-94ad-bfefc0b1de82 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7faf65a49e60> ====================
20/05/10 00:08:46 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589069325.69_b28fcdb0-299e-4cbd-a214-2614cb5607c8
20/05/10 00:08:46 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589069325.69_b28fcdb0-299e-4cbd-a214-2614cb5607c8
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/10 00:08:46 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/10 00:08:46 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/10 00:08:46 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589069325.69_b28fcdb0-299e-4cbd-a214-2614cb5607c8 on Spark master local
20/05/10 00:08:46 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/10 00:08:46 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589069325.69_b28fcdb0-299e-4cbd-a214-2614cb5607c8: Pipeline translated successfully. Computing outputs
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/10 00:08:47 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/10 00:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/10 00:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:37159
20/05/10 00:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/10 00:08:47 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/10 00:08:47 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589069325.69_b28fcdb0-299e-4cbd-a214-2614cb5607c8', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/10 00:08:47 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589069325.69', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:46695', 'job_port': u'0'}
20/05/10 00:08:47 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/10 00:08:47 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:45237.
20/05/10 00:08:47 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/10 00:08:47 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/10 00:08:47 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:44937.
20/05/10 00:08:47 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/10 00:08:47 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/10 00:08:47 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589069327.550204184","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589069327.550199294","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589069327.550197702","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589069327.550189040","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589069327.550165960","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589069327.550204184","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589069327.550199294","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589069327.550197702","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589069327.550189040","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589069327.550165960","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/10 00:08:47 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:45513
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/10 00:08:47 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/10 00:08:47 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589069325.69_b28fcdb0-299e-4cbd-a214-2614cb5607c8 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589069269.92_406272c7-af9c-4581-88a8-da94bf09eee4 failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 91.621s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 4m 21s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/ocjqkkocbvyec

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3136

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3136/display/redirect?page=changes>

Changes:

[chambers] Add a BeamJava test with Flatten with different input and output Coders

[jferriero] check lastAttemptedOffset non-null


------------------------------------------
[...truncated 567.54 KB...]
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
20/05/09 18:09:13 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/09 18:09:13 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/09 18:09:13 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7fa71cf3ee60> ====================
ERROR:grpc._channel:Exception iterating requests!
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 195, in consume_request_iterator
    request = next(request_iterator)
  File "apache_beam/runners/portability/artifact_service.py", line 317, in __next__
    raise self._queue.get()
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNIMPLEMENTED
	details = "Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService"
	debug_error_string = "{"created":"@1589047753.526406346","description":"Error received from peer ipv4:127.0.0.1:44755","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService","grpc_status":12}"
>
20/05/09 18:09:13 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589047753.07_bf793a81-ceae-437b-b5ae-c2d08e33c72d
20/05/09 18:09:13 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589047753.07_bf793a81-ceae-437b-b5ae-c2d08e33c72d
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/09 18:09:13 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/09 18:09:13 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/09 18:09:13 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589047753.07_bf793a81-ceae-437b-b5ae-c2d08e33c72d on Spark master local
20/05/09 18:09:13 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/09 18:09:13 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589047753.07_bf793a81-ceae-437b-b5ae-c2d08e33c72d: Pipeline translated successfully. Computing outputs
20/05/09 18:09:14 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/09 18:09:14 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/09 18:09:14 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:44387
20/05/09 18:09:14 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/09 18:09:14 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/09 18:09:14 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589047753.07_bf793a81-ceae-437b-b5ae-c2d08e33c72d', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/09 18:09:14 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589047753.07', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:44991', 'job_port': u'0'}
20/05/09 18:09:14 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/09 18:09:14 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:34547.
20/05/09 18:09:14 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/09 18:09:14 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/09 18:09:14 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/09 18:09:14 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/09 18:09:14 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:33007.
20/05/09 18:09:14 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/09 18:09:14 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/09 18:09:14 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589047754.473869097","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589047754.473864240","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589047754.473862886","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589047754.473854230","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589047754.473832948","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589047754.473869097","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589047754.473864240","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589047754.473862886","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589047754.473854230","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589047754.473832948","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/09 18:09:14 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:37609
20/05/09 18:09:14 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/09 18:09:14 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/09 18:09:14 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/09 18:09:14 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/09 18:09:14 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589047753.07_bf793a81-ceae-437b-b5ae-c2d08e33c72d finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7fa71cf3ee60> ====================
20/05/09 18:09:15 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589047754.91_4460ce86-47e3-49b1-8a5b-9c4f5776b627
20/05/09 18:09:15 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589047754.91_4460ce86-47e3-49b1-8a5b-9c4f5776b627
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/09 18:09:15 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/09 18:09:15 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/09 18:09:15 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/09 18:09:15 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/09 18:09:15 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/09 18:09:15 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589047754.91_4460ce86-47e3-49b1-8a5b-9c4f5776b627 on Spark master local
20/05/09 18:09:15 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/09 18:09:15 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589047754.91_4460ce86-47e3-49b1-8a5b-9c4f5776b627: Pipeline translated successfully. Computing outputs
20/05/09 18:09:16 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/09 18:09:16 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/09 18:09:16 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/09 18:09:16 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/09 18:09:16 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:34367
20/05/09 18:09:16 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589047754.91_4460ce86-47e3-49b1-8a5b-9c4f5776b627', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/09 18:09:16 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589047754.91', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:44991', 'job_port': u'0'}
20/05/09 18:09:16 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/09 18:09:16 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:40531.
20/05/09 18:09:16 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/09 18:09:16 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/09 18:09:16 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/09 18:09:16 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/09 18:09:16 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:35899.
20/05/09 18:09:16 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/09 18:09:16 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/09 18:09:16 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589047756.769238801","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589047756.769233578","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589047756.769231880","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589047756.769220405","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589047756.769165285","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589047756.769238801","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589047756.769233578","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589047756.769231880","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589047756.769220405","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589047756.769165285","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/09 18:09:16 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:43827
20/05/09 18:09:16 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/09 18:09:16 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/09 18:09:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/09 18:09:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/09 18:09:17 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/09 18:09:17 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589047754.91_4460ce86-47e3-49b1-8a5b-9c4f5776b627 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589047703.08_1dea8f47-4276-4aa8-9181-8c0d7252e9cb failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 82.114s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 4m 0s
64 actionable tasks: 50 executed, 14 from cache

Publishing build scan...
https://gradle.com/s/aacdbbzuvadok

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #3135

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/3135/display/redirect>

Changes:


------------------------------------------
[...truncated 565.94 KB...]
	at org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
	... 3 more
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 3: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 245, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 302, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 471, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 506, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 913, in process_bundle
    element.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 332, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 195, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 764, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 975, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 712, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 819, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssWARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f65228a9e60> ====================
20/05/09 12:06:57 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1589026017.46_17a2fcce-ab16-4c67-b0b9-27a5809d1360
20/05/09 12:06:57 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1589026017.46_17a2fcce-ab16-4c67-b0b9-27a5809d1360
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/09 12:06:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/09 12:06:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/09 12:06:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1589026017.46_17a2fcce-ab16-4c67-b0b9-27a5809d1360 on Spark master local
20/05/09 12:06:58 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/09 12:06:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589026017.46_17a2fcce-ab16-4c67-b0b9-27a5809d1360: Pipeline translated successfully. Computing outputs
20/05/09 12:06:58 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/09 12:06:58 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/09 12:06:58 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/09 12:06:58 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/09 12:06:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/09 12:06:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:33515
20/05/09 12:06:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/09 12:06:58 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/09 12:06:58 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1589026017.46_17a2fcce-ab16-4c67-b0b9-27a5809d1360', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/05/09 12:06:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1589026017.46', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:55299', 'job_port': u'0'}
20/05/09 12:06:58 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/09 12:06:58 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:41771.
20/05/09 12:06:58 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/09 12:06:58 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/09 12:06:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 29-1
20/05/09 12:06:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-2
20/05/09 12:06:58 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:34947.
20/05/09 12:06:58 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/09 12:06:58 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/09 12:06:58 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589026018.870912797","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589026018.870909147","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589026018.870908027","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589026018.870901881","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589026018.870886172","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589026018.870912797","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589026018.870909147","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589026018.870908027","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589026018.870901881","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589026018.870886172","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/09 12:06:58 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:46645
20/05/09 12:06:58 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/09 12:06:59 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-3
20/05/09 12:06:59 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-4
20/05/09 12:06:59 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 29-5
20/05/09 12:06:59 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1589026017.46_17a2fcce-ab16-4c67-b0b9-27a5809d1360 finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f65228a9e60> ====================
20/05/09 12:07:00 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1589026019.39_84d16321-a0ec-4320-9b2c-9a6c28a4f26a
20/05/09 12:07:00 INFO org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation test_windowing_1589026019.39_84d16321-a0ec-4320-9b2c-9a6c28a4f26a
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/05/09 12:07:00 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/05/09 12:07:00 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 7 files. (Enable logging at DEBUG level to see which files will be staged.)
20/05/09 12:07:00 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1589026019.39_84d16321-a0ec-4320-9b2c-9a6c28a4f26a on Spark master local
20/05/09 12:07:00 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/05/09 12:07:00 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589026019.39_84d16321-a0ec-4320-9b2c-9a6c28a4f26a: Pipeline translated successfully. Computing outputs
20/05/09 12:07:00 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"
capabilities: "beam:version:sdk_base:apache/beam_python2.7_sdk:2.22.0.dev"

20/05/09 12:07:00 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/05/09 12:07:00 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/05/09 12:07:00 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/05/09 12:07:00 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/05/09 12:07:00 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:42433
20/05/09 12:07:00 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/05/09 12:07:00 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/05/09 12:07:00 WARN apache_beam/options/pipeline_options.py:309: Discarding unparseable args: [u'--app_name=test_windowing_1589026019.39_84d16321-a0ec-4320-9b2c-9a6c28a4f26a', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=30', u'--enable_spark_metric_sinks'] 
20/05/09 12:07:00 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1589026019.39', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:55299', 'job_port': u'0'}
20/05/09 12:07:00 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/05/09 12:07:00 INFO apache_beam/runners/worker/sdk_worker.py:148: Creating insecure control channel for localhost:42707.
20/05/09 12:07:00 INFO apache_beam/runners/worker/sdk_worker.py:156: Control channel established.
20/05/09 12:07:00 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 30-1
20/05/09 12:07:00 INFO apache_beam/runners/worker/sdk_worker.py:195: Initializing SDKHarness with unbounded number of workers.
20/05/09 12:07:00 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-2
20/05/09 12:07:00 INFO apache_beam/runners/worker/sdk_worker.py:702: Creating insecure state channel for localhost:44047.
20/05/09 12:07:00 INFO apache_beam/runners/worker/sdk_worker.py:709: State channel established.
20/05/09 12:07:00 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for 
20/05/09 12:07:00 ERROR apache_beam/runners/worker/data_plane.py:535: Failed to read inputs in the data plane.
Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589026020.983213589","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589026020.983208752","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589026020.983207166","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589026020.983198989","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589026020.983173065","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
> Traceback (most recent call last):
  File "apache_beam/runners/worker/data_plane.py", line 528, in _read_inputs
    for elements in elements_iterator:
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 413, in next
    return self._next()
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/build/gradleenv/1866363813/local/lib/python2.7/site-packages/grpc/_channel.py",> line 689, in _next
    raise self
_MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNAVAILABLE
	details = "DNS resolution failed"
	debug_error_string = "{"created":"@1589026020.983213589","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":3981,"referenced_errors":[{"created":"@1589026020.983208752","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/resolving_lb_policy.cc","file_line":214,"referenced_errors":[{"created":"@1589026020.983207166","description":"DNS resolution failed","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":357,"grpc_status":14,"referenced_errors":[{"created":"@1589026020.983198989","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244,"referenced_errors":[{"created":"@1589026020.983173065","description":"C-ares status is not ARES_SUCCESS: Misformatted domain name","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":244}]}]}]}]}"
>

20/05/09 12:07:00 INFO apache_beam/runners/worker/data_plane.py:634: Creating client data channel for localhost:39603
20/05/09 12:07:00 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/05/09 12:07:01 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-3
20/05/09 12:07:01 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-4
20/05/09 12:07:01 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-5
20/05/09 12:07:01 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: getProcessBundleDescriptor request with id 30-6
20/05/09 12:07:01 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1589026019.39_84d16321-a0ec-4320-9b2c-9a6c28a4f26a finished.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_multimap_multiside_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner/fn_runner_test.py", line 259, in test_multimap_multiside_input
    equal_to([('a', [1, 3], [1, 2, 3]), ('b', [2], [1, 2, 3])]))
  File "apache_beam/pipeline.py", line 550, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline test_multimap_multiside_input_1589025971.19_ae0f8ea5-0180-43e7-8453-8bce9fe3fd7d failed in state FAILED: java.lang.IllegalArgumentException: Multiple entries with same key: ref_PCollection_PCollection_20=(Broadcast(37),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder)) and ref_PCollection_PCollection_20=(Broadcast(36),WindowedValue$FullWindowedValueCoder(KvCoder(ByteArrayCoder,VarLongCoder),GlobalWindow$Coder))

----------------------------------------------------------------------
Ran 40 tests in 77.597s

FAILED (errors=1, skipped=11)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 213

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3m 29s
64 actionable tasks: 49 executed, 15 from cache

Publishing build scan...
https://gradle.com/s/54ys5gaxznnma

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org