You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/03/05 06:18:02 UTC

Build failed in Jenkins: beam_PostCommit_Python_VR_Spark #2543

See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/2543/display/redirect>

Changes:


------------------------------------------
[...truncated 464.03 KB...]
	at java.lang.Thread.run(Thread.java:748)
ERROR:root:java.lang.RuntimeException: Error received from SDK harness for instruction 6: Traceback (most recent call last):
  File "apache_beam/runners/worker/sdk_worker.py", line 190, in _execute
    response = task()
  File "apache_beam/runners/worker/sdk_worker.py", line 247, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File "apache_beam/runners/worker/sdk_worker.py", line 416, in do_instruction
    getattr(request, request_type), request.instruction_id)
  File "apache_beam/runners/worker/sdk_worker.py", line 451, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File "apache_beam/runners/worker/bundle_processor.py", line 842, in process_bundle
    input_op_by_transform_id[data.transform_id].process_encoded(data.data)
  File "apache_beam/runners/worker/bundle_processor.py", line 217, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 334, in output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 197, in receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 793, in process
    delayed_application = self.dofn_runner.process_with_sized_restriction(o)
  File "apache_beam/runners/common.py", line 968, in process_with_sized_restriction
    watermark_estimator=watermark_estimator)
  File "apache_beam/runners/common.py", line 710, in invoke_process
    windowed_value, additional_args, additional_kwargs)
  File "apache_beam/runners/common.py", line 812, in _invoke_process_per_window
    self.threadsafe_restriction_tracker.check_done()
  File "apache_beam/runners/sdf_utils.py", line 115, in check_done
    return self._restriction_tracker.check_done()
  File "apache_beam/io/restriction_trackers.py", line 101, in check_done
    self._range.stop))
ValueError: OffsetRestrictionTracker is not done since work in range [0, 6) has not been claimed.

INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
.sssINFO:apache_beam.runners.portability.fn_api_runner_transforms:==================== <function lift_combiners at 0x7f9854854578> ====================
20/03/05 06:17:54 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowed_pardo_state_timers_1583389073.75_ab7edc99-3afa-4a47-8ae8-678970ade8d9
20/03/05 06:17:54 INFO org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation: Starting job invocation test_windowed_pardo_state_timers_1583389073.75_ab7edc99-3afa-4a47-8ae8-678970ade8d9
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/03/05 06:17:55 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/03/05 06:17:55 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 1 files. (Enable logging at DEBUG level to see which files will be staged.)
20/03/05 06:17:55 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowed_pardo_state_timers_1583389073.75_ab7edc99-3afa-4a47-8ae8-678970ade8d9 on Spark master local
20/03/05 06:17:55 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/03/05 06:17:55 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1583389073.75_ab7edc99-3afa-4a47-8ae8-678970ade8d9: Pipeline translated successfully. Computing outputs
20/03/05 06:17:55 INFO org.apache.beam.runners.fnexecution.artifact.AbstractArtifactRetrievalService: GetManifest for __no_artifacts_staged__
20/03/05 06:17:55 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"

20/03/05 06:17:55 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/03/05 06:17:55 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/03/05 06:17:55 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/03/05 06:17:55 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/03/05 06:17:55 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:38943
20/03/05 06:17:55 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/03/05 06:17:55 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/03/05 06:17:55 WARN apache_beam/options/pipeline_options.py:290: Discarding unparseable args: [u'--app_name=test_windowed_pardo_state_timers_1583389073.75_ab7edc99-3afa-4a47-8ae8-678970ade8d9', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=28', u'--enable_spark_metric_sinks'] 
20/03/05 06:17:55 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowed_pardo_state_timers_1583389073.75', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:47271', 'job_port': u'0'}
20/03/05 06:17:55 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/03/05 06:17:55 INFO apache_beam/runners/worker/sdk_worker.py:98: Creating insecure control channel for localhost:37577.
20/03/05 06:17:55 INFO apache_beam/runners/worker/sdk_worker.py:106: Control channel established.
20/03/05 06:17:55 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 27-1
20/03/05 06:17:55 INFO apache_beam/runners/worker/sdk_worker.py:141: Initializing SDKHarness with unbounded number of workers.
20/03/05 06:17:55 INFO apache_beam/runners/worker/sdk_worker.py:627: Creating insecure state channel for localhost:44897.
20/03/05 06:17:55 INFO apache_beam/runners/worker/sdk_worker.py:634: State channel established.
20/03/05 06:17:55 INFO apache_beam/runners/worker/data_plane.py:533: Creating client data channel for localhost:43723
20/03/05 06:17:55 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/03/05 06:17:56 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowed_pardo_state_timers_1583389073.75_ab7edc99-3afa-4a47-8ae8-678970ade8d9 finished.
20/03/05 06:17:56 WARN org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService: Failed to remove job staging directory for token {"sessionId":"job_594a54e5-31bc-4d17-acb2-e452234b5353","basePath":"/tmp/sparktestuGhrjR"}: {}
java.io.FileNotFoundException: /tmp/sparktestuGhrjR/job_594a54e5-31bc-4d17-acb2-e452234b5353/MANIFEST (No such file or directory)
	at java.io.FileInputStream.open0(Native Method)
	at java.io.FileInputStream.open(FileInputStream.java:195)
	at java.io.FileInputStream.<init>(FileInputStream.java:138)
	at org.apache.beam.sdk.io.LocalFileSystem.open(LocalFileSystem.java:118)
	at org.apache.beam.sdk.io.LocalFileSystem.open(LocalFileSystem.java:82)
	at org.apache.beam.sdk.io.FileSystems.open(FileSystems.java:252)
	at org.apache.beam.runners.fnexecution.artifact.BeamFileSystemArtifactRetrievalService.loadManifest(BeamFileSystemArtifactRetrievalService.java:88)
	at org.apache.beam.runners.fnexecution.artifact.BeamFileSystemArtifactStagingService.removeArtifacts(BeamFileSystemArtifactStagingService.java:92)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.lambda$createJobService$0(JobServerDriver.java:63)
	at org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService.lambda$run$0(InMemoryJobService.java:245)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.setState(JobInvocation.java:246)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.access$200(JobInvocation.java:48)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation$1.onSuccess(JobInvocation.java:112)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation$1.onSuccess(JobInvocation.java:98)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1058)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.INFO:apache_beam.runners.portability.fn_api_runner_transforms:==================== <function lift_combiners at 0x7f9854854578> ====================
20/03/05 06:17:57 INFO org.apache.beam.runners.spark.SparkJobInvoker: Invoking job test_windowing_1583389076.42_23ee6375-3333-4eea-b40a-163e599b3f9d
20/03/05 06:17:57 INFO org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation: Starting job invocation test_windowing_1583389076.42_23ee6375-3333-4eea-b40a-163e599b3f9d
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
20/03/05 06:17:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath
20/03/05 06:17:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 1 files. (Enable logging at DEBUG level to see which files will be staged.)
20/03/05 06:17:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Running job test_windowing_1583389076.42_23ee6375-3333-4eea-b40a-163e599b3f9d on Spark master local
20/03/05 06:17:58 WARN org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions: Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not consistent with equals. That might cause issues on some runners.
20/03/05 06:17:58 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1583389076.42_23ee6375-3333-4eea-b40a-163e599b3f9d: Pipeline translated successfully. Computing outputs
20/03/05 06:17:58 INFO org.apache.beam.runners.fnexecution.artifact.AbstractArtifactRetrievalService: GetManifest for __no_artifacts_staged__
20/03/05 06:17:58 INFO org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory: Closing environment urn: "beam:env:process:v1"
payload: "\032\202\001<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh">
capabilities: "beam:coder:varint:v1"
capabilities: "beam:coder:bytes:v1"
capabilities: "beam:coder:timer:v1"
capabilities: "beam:coder:global_window:v1"
capabilities: "beam:coder:interval_window:v1"
capabilities: "beam:coder:iterable:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:windowed_value:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:string_utf8:v1"
capabilities: "beam:coder:length_prefix:v1"
capabilities: "beam:coder:bool:v1"
capabilities: "beam:coder:kv:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:protocol:progress_reporting:v0"
capabilities: "beam:protocol:worker_status:v1"

20/03/05 06:17:58 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: 1 Beam Fn Logging clients still connected during shutdown.
20/03/05 06:17:58 WARN org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer: Hanged up for unknown endpoint.
20/03/05 06:17:58 INFO org.apache.beam.runners.fnexecution.logging.GrpcLoggingService: Beam Fn Logging client connected.
20/03/05 06:17:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:98: Logging handler created.
20/03/05 06:17:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:79: Status HTTP server running at localhost:32963
20/03/05 06:17:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:125: semi_persistent_directory: /tmp
20/03/05 06:17:58 WARN <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:240: No session file found: /tmp/staged/pickled_main_session. Functions defined in __main__ (interactive session) may fail. 
20/03/05 06:17:58 WARN apache_beam/options/pipeline_options.py:290: Discarding unparseable args: [u'--app_name=test_windowing_1583389076.42_23ee6375-3333-4eea-b40a-163e599b3f9d', u'--job_server_timeout=60', u'--pipeline_type_check', u'--direct_runner_use_stacked_bundle', u'--options_id=29', u'--enable_spark_metric_sinks'] 
20/03/05 06:17:58 INFO <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/worker/sdk_worker_main.py>:138: Python sdk harness started with pipeline_options: {'runner': u'None', 'experiments': [u'beam_fn_api'], 'environment_cache_millis': u'10000', 'artifact_port': u'0', 'environment_type': u'PROCESS', 'sdk_location': u'container', 'job_name': u'test_windowing_1583389076.42', 'environment_config': u'{"command": "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build/sdk_worker.sh"}',> 'expansion_port': u'0', 'spark_master_url': u'local', 'sdk_worker_parallelism': u'1', 'job_endpoint': u'localhost:47271', 'job_port': u'0'}
20/03/05 06:17:58 INFO apache_beam/runners/worker/statecache.py:154: Creating state cache with size 0
20/03/05 06:17:58 INFO apache_beam/runners/worker/sdk_worker.py:98: Creating insecure control channel for localhost:45877.
20/03/05 06:17:58 INFO apache_beam/runners/worker/sdk_worker.py:106: Control channel established.
20/03/05 06:17:58 INFO org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService: Beam Fn Control client connected with id 28-1
20/03/05 06:17:58 INFO apache_beam/runners/worker/sdk_worker.py:141: Initializing SDKHarness with unbounded number of workers.
20/03/05 06:17:58 INFO apache_beam/runners/worker/sdk_worker.py:627: Creating insecure state channel for localhost:40387.
20/03/05 06:17:58 INFO apache_beam/runners/worker/sdk_worker.py:634: State channel established.
20/03/05 06:17:58 INFO apache_beam/runners/worker/data_plane.py:533: Creating client data channel for localhost:41985
20/03/05 06:17:58 INFO org.apache.beam.runners.fnexecution.data.GrpcDataService: Beam Fn Data client connected.
20/03/05 06:17:59 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Job test_windowing_1583389076.42_23ee6375-3333-4eea-b40a-163e599b3f9d finished.
20/03/05 06:17:59 WARN org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService: Failed to remove job staging directory for token {"sessionId":"job_b18bfdcf-c387-48dd-af3d-4a0ece5b6041","basePath":"/tmp/sparktestuGhrjR"}: {}
java.io.FileNotFoundException: /tmp/sparktestuGhrjR/job_b18bfdcf-c387-48dd-af3d-4a0ece5b6041/MANIFEST (No such file or directory)
	at java.io.FileInputStream.open0(Native Method)
	at java.io.FileInputStream.open(FileInputStream.java:195)
	at java.io.FileInputStream.<init>(FileInputStream.java:138)
	at org.apache.beam.sdk.io.LocalFileSystem.open(LocalFileSystem.java:118)
	at org.apache.beam.sdk.io.LocalFileSystem.open(LocalFileSystem.java:82)
	at org.apache.beam.sdk.io.FileSystems.open(FileSystems.java:252)
	at org.apache.beam.runners.fnexecution.artifact.BeamFileSystemArtifactRetrievalService.loadManifest(BeamFileSystemArtifactRetrievalService.java:88)
	at org.apache.beam.runners.fnexecution.artifact.BeamFileSystemArtifactStagingService.removeArtifacts(BeamFileSystemArtifactStagingService.java:92)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.lambda$createJobService$0(JobServerDriver.java:63)
	at org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService.lambda$run$0(InMemoryJobService.java:245)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.setState(JobInvocation.java:246)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.access$200(JobInvocation.java:48)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation$1.onSuccess(JobInvocation.java:112)
	at org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation$1.onSuccess(JobInvocation.java:98)
	at org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1058)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
.
======================================================================
ERROR: test_flatten_same_pcollections (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "apache_beam/runners/portability/fn_api_runner_test.py", line 577, in test_flatten_same_pcollections
    assert_that((pc, pc, pc) | beam.Flatten(), equal_to(['a', 'b'] * 3))
  File "apache_beam/pipeline.py", line 522, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/runners/portability/portable_runner.py", line 532, in wait_until_finish
    (self._job_id, self._state, self._last_error_message()))
RuntimeError: Pipeline test_flatten_same_pcollections_1583388993.56_26b2e395-7fe2-40f2-9806-6319d0310926 failed in state FAILED: java.io.IOException: Failed to create local dir in /tmp/blockmgr-2259f8f5-2984-4599-b168-04ee174bb6fc/23.

======================================================================
ERROR: test_flattened_side_input (__main__.SparkRunnerTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/apache_beam/runners/portability/spark_runner_test.py",> line 151, in test_flattened_side_input
    self).test_flattened_side_input(with_transcoding=False)
  File "apache_beam/runners/portability/fn_api_runner_test.py", line 214, in test_flattened_side_input
    label='CheckFlattenOfSideInput')
  File "apache_beam/pipeline.py", line 522, in __exit__
    self.run().wait_until_finish()
  File "apache_beam/pipeline.py", line 495, in run
    self._options).run(False)
  File "apache_beam/pipeline.py", line 508, in run
    return self.runner.run_pipeline(self, self._options)
  File "apache_beam/runners/portability/portable_runner.py", line 388, in run_pipeline
    job_service_handle.submit(proto_pipeline)
  File "apache_beam/runners/portability/portable_runner.py", line 105, in submit
    prepare_response.staging_session_token)
  File "apache_beam/runners/portability/portable_runner.py", line 190, in stage
    staging_location='')
  File "apache_beam/runners/portability/stager.py", line 150, in stage_job_resources
    temp_dir = temp_dir or tempfile.mkdtemp()
  File "/usr/lib/python2.7/tempfile.py", line 339, in mkdtemp
    _os.mkdir(file, 0700)
OSError: [Errno 28] No space left on device: '/tmp/tmpg73PNI'

----------------------------------------------------------------------
Ran 39 tests in 112.137s

FAILED (errors=2, skipped=10)

> Task :sdks:python:test-suites:portable:py2:sparkValidatesRunner FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/ws/src/sdks/python/test-suites/portable/py2/build.gradle'> line: 248

* What went wrong:
Execution failed for task ':sdks:python:test-suites:portable:py2:sparkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 4m 43s
62 actionable tasks: 48 executed, 14 from cache

Publishing build scan...
https://gradle.com/s/wikgriy42fnrk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Python_VR_Spark #2544

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_VR_Spark/2544/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org