You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/10/01 19:02:55 UTC

Build failed in Jenkins: beam_PostCommit_Python2 #601

See <https://builds.apache.org/job/beam_PostCommit_Python2/601/display/redirect>

Changes:


------------------------------------------
[...truncated 1.08 MB...]
Exception: (u"BigQuery jobs failed. BQ error: %s [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs']", <JobStatus
 errorResult: <ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>
 errors: [<ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>]
 state: u'DONE'>)

root: INFO: 2019-10-01T18:22:19.911Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/WaitForDestinationLoadJobs/WaitForDestinationLoadJobs+WriteWithMultipleDests/BigQueryBatchFileLoads/WaitForTempTableLoadJobs/WaitForTempTableLoadJobs+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(TriggerCopyJobs)/ParDo(TriggerCopyJobs)
root: INFO: 2019-10-01T18:22:19.986Z: JOB_MESSAGE_DEBUG: Value "WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(TriggerCopyJobs).out" materialized.
root: INFO: 2019-10-01T18:22:20.060Z: JOB_MESSAGE_BASIC: Executing operation WriteWithMultipleDests/BigQueryBatchFileLoads/WaitForCopyJobs/_UnpickledSideInput(ParDo(TriggerCopyJobs).out.0)
root: INFO: 2019-10-01T18:22:20.113Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/WaitForCopyJobs/_UnpickledSideInput(ParDo(TriggerCopyJobs).out.0)
root: INFO: 2019-10-01T18:22:20.177Z: JOB_MESSAGE_DEBUG: Value "WriteWithMultipleDests/BigQueryBatchFileLoads/WaitForCopyJobs/_UnpickledSideInput(ParDo(TriggerCopyJobs).out.0).output" materialized.
root: INFO: 2019-10-01T18:22:20.237Z: JOB_MESSAGE_BASIC: Executing operation WriteWithMultipleDests/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/PassTables/PassTables+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/AddUselessValue+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Reify+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Write
root: INFO: 2019-10-01T18:22:21.331Z: JOB_MESSAGE_ERROR: Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 649, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 176, in execute
    op.start()
  File "dataflow_worker/native_operations.py", line 38, in dataflow_worker.native_operations.NativeReadOperation.start
    def start(self):
  File "dataflow_worker/native_operations.py", line 39, in dataflow_worker.native_operations.NativeReadOperation.start
    with self.scoped_start_state:
  File "dataflow_worker/native_operations.py", line 44, in dataflow_worker.native_operations.NativeReadOperation.start
    with self.spec.source.reader() as reader:
  File "dataflow_worker/native_operations.py", line 54, in dataflow_worker.native_operations.NativeReadOperation.start
    self.output(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 256, in apache_beam.runners.worker.operations.Operation.output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 143, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 593, in apache_beam.runners.worker.operations.DoOperation.process
    with self.scoped_process_state:
  File "apache_beam/runners/worker/operations.py", line 594, in apache_beam.runners.worker.operations.DoOperation.process
    delayed_application = self.dofn_receiver.receive(o)
  File "apache_beam/runners/common.py", line 776, in apache_beam.runners.common.DoFnRunner.receive
    self.process(windowed_value)
  File "apache_beam/runners/common.py", line 782, in apache_beam.runners.common.DoFnRunner.process
    self._reraise_augmented(exn)
  File "apache_beam/runners/common.py", line 849, in apache_beam.runners.common.DoFnRunner._reraise_augmented
    raise_with_traceback(new_exn)
  File "apache_beam/runners/common.py", line 780, in apache_beam.runners.common.DoFnRunner.process
    return self.do_fn_invoker.invoke_process(windowed_value)
  File "apache_beam/runners/common.py", line 587, in apache_beam.runners.common.PerWindowInvoker.invoke_process
    self._invoke_process_per_window(
  File "apache_beam/runners/common.py", line 660, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
    windowed_value, self.process_method(*args_for_process))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 511, in process
    'BigQuery jobs failed. BQ error: %s', self._latest_error)
Exception: (u"BigQuery jobs failed. BQ error: %s [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs']", <JobStatus
 errorResult: <ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>
 errors: [<ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>]
 state: u'DONE'>)

root: INFO: 2019-10-01T18:22:24.480Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/PassTables/PassTables+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/AddUselessValue+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Reify+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Write
root: INFO: 2019-10-01T18:22:24.567Z: JOB_MESSAGE_BASIC: Executing operation WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Close
root: INFO: 2019-10-01T18:22:24.623Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Close
root: INFO: 2019-10-01T18:22:24.709Z: JOB_MESSAGE_BASIC: Executing operation WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/GroupByWindow+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/GetTableNames+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/Delete
root: INFO: 2019-10-01T18:22:25.023Z: JOB_MESSAGE_ERROR: Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 649, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 176, in execute
    op.start()
  File "dataflow_worker/native_operations.py", line 38, in dataflow_worker.native_operations.NativeReadOperation.start
    def start(self):
  File "dataflow_worker/native_operations.py", line 39, in dataflow_worker.native_operations.NativeReadOperation.start
    with self.scoped_start_state:
  File "dataflow_worker/native_operations.py", line 44, in dataflow_worker.native_operations.NativeReadOperation.start
    with self.spec.source.reader() as reader:
  File "dataflow_worker/native_operations.py", line 54, in dataflow_worker.native_operations.NativeReadOperation.start
    self.output(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 256, in apache_beam.runners.worker.operations.Operation.output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 143, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 593, in apache_beam.runners.worker.operations.DoOperation.process
    with self.scoped_process_state:
  File "apache_beam/runners/worker/operations.py", line 594, in apache_beam.runners.worker.operations.DoOperation.process
    delayed_application = self.dofn_receiver.receive(o)
  File "apache_beam/runners/common.py", line 776, in apache_beam.runners.common.DoFnRunner.receive
    self.process(windowed_value)
  File "apache_beam/runners/common.py", line 782, in apache_beam.runners.common.DoFnRunner.process
    self._reraise_augmented(exn)
  File "apache_beam/runners/common.py", line 849, in apache_beam.runners.common.DoFnRunner._reraise_augmented
    raise_with_traceback(new_exn)
  File "apache_beam/runners/common.py", line 780, in apache_beam.runners.common.DoFnRunner.process
    return self.do_fn_invoker.invoke_process(windowed_value)
  File "apache_beam/runners/common.py", line 587, in apache_beam.runners.common.PerWindowInvoker.invoke_process
    self._invoke_process_per_window(
  File "apache_beam/runners/common.py", line 660, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
    windowed_value, self.process_method(*args_for_process))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 511, in process
    'BigQuery jobs failed. BQ error: %s', self._latest_error)
Exception: (u"BigQuery jobs failed. BQ error: %s [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs']", <JobStatus
 errorResult: <ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>
 errors: [<ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>]
 state: u'DONE'>)

root: INFO: 2019-10-01T18:22:26.908Z: JOB_MESSAGE_ERROR: Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 649, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 176, in execute
    op.start()
  File "dataflow_worker/native_operations.py", line 38, in dataflow_worker.native_operations.NativeReadOperation.start
    def start(self):
  File "dataflow_worker/native_operations.py", line 39, in dataflow_worker.native_operations.NativeReadOperation.start
    with self.scoped_start_state:
  File "dataflow_worker/native_operations.py", line 44, in dataflow_worker.native_operations.NativeReadOperation.start
    with self.spec.source.reader() as reader:
  File "dataflow_worker/native_operations.py", line 54, in dataflow_worker.native_operations.NativeReadOperation.start
    self.output(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 256, in apache_beam.runners.worker.operations.Operation.output
    cython.cast(Receiver, self.receivers[output_index]).receive(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 143, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive
    self.consumer.process(windowed_value)
  File "apache_beam/runners/worker/operations.py", line 593, in apache_beam.runners.worker.operations.DoOperation.process
    with self.scoped_process_state:
  File "apache_beam/runners/worker/operations.py", line 594, in apache_beam.runners.worker.operations.DoOperation.process
    delayed_application = self.dofn_receiver.receive(o)
  File "apache_beam/runners/common.py", line 776, in apache_beam.runners.common.DoFnRunner.receive
    self.process(windowed_value)
  File "apache_beam/runners/common.py", line 782, in apache_beam.runners.common.DoFnRunner.process
    self._reraise_augmented(exn)
  File "apache_beam/runners/common.py", line 849, in apache_beam.runners.common.DoFnRunner._reraise_augmented
    raise_with_traceback(new_exn)
  File "apache_beam/runners/common.py", line 780, in apache_beam.runners.common.DoFnRunner.process
    return self.do_fn_invoker.invoke_process(windowed_value)
  File "apache_beam/runners/common.py", line 587, in apache_beam.runners.common.PerWindowInvoker.invoke_process
    self._invoke_process_per_window(
  File "apache_beam/runners/common.py", line 660, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
    windowed_value, self.process_method(*args_for_process))
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 511, in process
    'BigQuery jobs failed. BQ error: %s', self._latest_error)
Exception: (u"BigQuery jobs failed. BQ error: %s [while running 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs']", <JobStatus
 errorResult: <ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>
 errors: [<ErrorProto
 message: u'Not found: Table apache-beam-testing:python_bq_file_loads_1569953797522.beam_load_2019_10_01_182126_28_26468470b703e8f653c82cf01ae6f1e2_72d4c95d2f0a4ebe83b4a5bc03dae41f was not found in location US'
 reason: u'notFound'>]
 state: u'DONE'>)

root: INFO: 2019-10-01T18:22:26.947Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/PassTables/PassTables+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/AddUselessValue+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Reify+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Write
root: INFO: 2019-10-01T18:22:27.026Z: JOB_MESSAGE_DEBUG: Executing failure step failure98
root: INFO: 2019-10-01T18:22:27.060Z: JOB_MESSAGE_ERROR: Workflow failed. Causes: S53:WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/WaitForCopyJobs/WaitForCopyJobs+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/PassTables/PassTables+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/AddUselessValue+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Reify+WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers: 
  beamapp-jenkins-100118164-10011116-zcm4-harness-f4z6
      Root cause: Work item failed.,
  beamapp-jenkins-100118164-10011116-zcm4-harness-f4z6
      Root cause: Work item failed.,
  beamapp-jenkins-100118164-10011116-zcm4-harness-f4z6
      Root cause: Work item failed.,
  beamapp-jenkins-100118164-10011116-zcm4-harness-f4z6
      Root cause: Work item failed.
root: INFO: 2019-10-01T18:22:27.172Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/DeduplicateTables/GroupByWindow+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/GetTableNames+WriteWithMultipleDests/BigQueryBatchFileLoads/RemoveTempTables/Delete
root: INFO: 2019-10-01T18:22:27.316Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2019-10-01T18:22:27.377Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown.
root: INFO: 2019-10-01T18:22:27.407Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2019-10-01T18:24:35.957Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-01T18:24:36.006Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2019-10-01T18:24:36.043Z: JOB_MESSAGE_DEBUG: Tearing down pending resources...
root: INFO: Job 2019-10-01_11_16_52-2126639192586986622 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: nosetests-postCommitIT-df.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 45 tests in 3675.379s

FAILED (SKIP=4, errors=1)
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_57-4052755215415788582?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_16_52-2126639192586986622?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_25_01-9216175690740839149?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_32_49-10251103466635280525?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_39_46-800259807984197400?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_54-12521119488105620990?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_21_21-14144697313749105006?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_39_09-13898229036334644267?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_52-14543124305917047571?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_22_22-251243316341746314?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_29_05-16787358592318294603?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_36_01-6705290148793551863?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_57-15519501043501032445?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_14_26-7882150120380362576?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_21_42-14053090978740289981?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_28_41-17745607315563646308?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_34_59-1724294451073617934?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_52-11761172600277380537?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_09_42-15783065206301195607?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_17_07-13036627972621982983?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_24_39-7567537103971007801?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_31_57-12164738417020577122?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_39_25-4786652535262295642?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_47_34-14726458103798234276?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_55_35-9937395425351708938?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_52-12429380921374637591?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_08_53-10212653044951782560?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_18_23-10414333269954534664?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_24_42-17885069953596570330?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_32_02-170512051935299755?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_52-6514905900843747786?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_10_51-17867019448795545215?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_19_15-14421570688313583824?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_26_40-1189494790065307?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_33_52-11265177504277443548?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_01_53-5623108700753501758?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_10_31-9972514633641227542?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_20_37-396752910431678684?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_28_23-4552635775422201750?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-01_11_35_42-6467036173305205401?project=apache-beam-testing

> Task :sdks:python:test-suites:dataflow:py2:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/test-suites/dataflow/py2/build.gradle'> line: 85

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py2:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 2m 34s
111 actionable tasks: 86 executed, 22 from cache, 3 up-to-date

Publishing build scan...
https://gradle.com/s/umvpgvzyyxyg6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Python2 #603

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python2/603/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python2 #602

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python2/602/display/redirect?page=changes>

Changes:

[bhulette] Always use UTC when creating instances of baselocal

[robertwb] [BEAM-6896] Loosen PyYAML dependency.


------------------------------------------
[...truncated 560.75 KB...]
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2) (75289372548343dc1658ab2b4d506854) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2).
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) switched from CREATED to DEPLOYING.
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak safety net for task GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) [DEPLOYING]
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) [DEPLOYING].
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Registering task at network: GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) [DEPLOYING].
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) switched from DEPLOYING to RUNNING.
[CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) (2/2) (5fa570871443111e640579424af2c9df) switched from RUNNING to FINISHED.
[CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) (2/2) (5fa570871443111e640579424af2c9df).
[CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (1/2) (b6d6efe85c007988b7cc57f7a3e9066f) switched from RUNNING to FINISHED.
[CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (1/2) (b6d6efe85c007988b7cc57f7a3e9066f).
[CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (2/2) (e111fe4a68e01b813a1553d944914ac4) switched from RUNNING to FINISHED.
[CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (2/2) (e111fe4a68e01b813a1553d944914ac4).
[CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) (2/2) (5fa570871443111e640579424af2c9df) [FINISHED]
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) 5fa570871443111e640579424af2c9df.
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - CHAIN MapPartition (MapPartition at [6]{Map(<lambda at external_test.py:377>), Map(<lambda at external_test.py:378>), assert_that}) -> FlatMap (FlatMap at ExtractOutput[0]) (2/2) (5fa570871443111e640579424af2c9df) switched from RUNNING to FINISHED.
[CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (2/2) (e111fe4a68e01b813a1553d944914ac4) [FINISHED]
[CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (1/2) (b6d6efe85c007988b7cc57f7a3e9066f) [FINISHED]
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) e111fe4a68e01b813a1553d944914ac4.
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) b6d6efe85c007988b7cc57f7a3e9066f.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (2/2) (e111fe4a68e01b813a1553d944914ac4) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) switched from CREATED to SCHEDULED.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) switched from SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (attempt #0) to 616be0f9-4fed-48bf-91bc-0a2a2fdd013b @ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - CHAIN Filter (UnionFixFilter) -> Map (Key Extractor) -> GroupCombine (GroupCombine at GroupCombine: assert_that/Group/GroupByKey) -> Map (Key Extractor) (1/2) (b6d6efe85c007988b7cc57f7a3e9066f) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2).
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) switched from CREATED to DEPLOYING.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak safety net for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) [DEPLOYING]
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) [DEPLOYING].
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Registering task at network: MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) [DEPLOYING].
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) switched from DEPLOYING to RUNNING.
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) switched from RUNNING to FINISHED.
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9).
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) [FINISHED]
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2) (75289372548343dc1658ab2b4d506854) switched from RUNNING to FINISHED.
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2) (75289372548343dc1658ab2b4d506854).
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task GroupReduce (GroupReduce at assert_that/Group/GroupByKey) 4d217066c58809636211d119dd9d8da9.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) switched from CREATED to SCHEDULED.
[GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2) (75289372548343dc1658ab2b4d506854) [FINISHED]
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task GroupReduce (GroupReduce at assert_that/Group/GroupByKey) 75289372548343dc1658ab2b4d506854.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) switched from SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (attempt #0) to 616be0f9-4fed-48bf-91bc-0a2a2fdd013b @ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (2/2) (4d217066c58809636211d119dd9d8da9) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - GroupReduce (GroupReduce at assert_that/Group/GroupByKey) (1/2) (75289372548343dc1658ab2b4d506854) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2).
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) switched from CREATED to DEPLOYING.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak safety net for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) [DEPLOYING]
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) [DEPLOYING].
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Registering task at network: MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) [DEPLOYING].
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) switched from DEPLOYING to RUNNING.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b).
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) switched from CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) [FINISHED]
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) 742b5f2543a447d82b3b40f74c78222b.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) switched from SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink (DiscardingOutput) (2/2) (attempt #0) to 616be0f9-4fed-48bf-91bc-0a2a2fdd013b @ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink (DiscardingOutput) (2/2).
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/2) (742b5f2543a447d82b3b40f74c78222b) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak safety net for task DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) [DEPLOYING]
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) [DEPLOYING].
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Registering task at network: DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) [DEPLOYING].
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) switched from DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f).
[DataSink (DiscardingOutput) (2/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) [FINISHED]
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task DataSink (DiscardingOutput) 0dbdb9d0d5560e9dae634630b651d24f.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (2/2) (0dbdb9d0d5560e9dae634630b651d24f) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0).
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) switched from CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) [FINISHED]
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) switched from SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) 08a683f24cd517519bf3a6383e0154e0.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink (DiscardingOutput) (1/2) (attempt #0) to 616be0f9-4fed-48bf-91bc-0a2a2fdd013b @ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/2) (08a683f24cd517519bf3a6383e0154e0) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink (DiscardingOutput) (1/2).
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak safety net for task DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) [DEPLOYING]
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) [DEPLOYING].
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Registering task at network: DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) [DEPLOYING].
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) switched from DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6).
[DataSink (DiscardingOutput) (1/2)] INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are closed for task DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) [FINISHED]
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and sending final execution state FINISHED to JobManager for task DataSink (DiscardingOutput) 71684e56b1d04b4d5dc71748706479f6.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink (DiscardingOutput) (1/2) (71684e56b1d04b4d5dc71748706479f6) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Job BeamApp-root-1001192150-5b26df41 (3dbeef5babb92f6da7b9b1fb1f28ea75) switched from state RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Job 3dbeef5babb92f6da7b9b1fb1f28ea75 reached globally terminal state FINISHED.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.jobmaster.JobMaster - Stopping the JobMaster for job BeamApp-root-1001192150-5b26df41(3dbeef5babb92f6da7b9b1fb1f28ea75).
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl - Suspending SlotPool.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.jobmaster.JobMaster - Close ResourceManager connection f08768a9b39ce5d7055b6bb021972c35: JobManager is shutting down..
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.slot.TaskSlotTable - Free slot TaskSlot(index:1, state:ACTIVE, resource profile: ResourceProfile{cpuCores=1.7976931348623157E308, heapMemoryInMB=2147483647, directMemoryInMB=2147483647, nativeMemoryInMB=2147483647, networkMemoryInMB=2147483647}, allocationId: b23522e5b06bb2d8911d6c1468c57857, jobId: 3dbeef5babb92f6da7b9b1fb1f28ea75).
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl - Stopping SlotPool.
[flink-akka.actor.default-dispatcher-7] INFO org.apache.flink.runtime.resourcemanager.StandaloneResourceManager - Disconnect job manager a5396978043c970d48b812e73d0a4ceb@akka://flink/user/jobmanager_1 for job 3dbeef5babb92f6da7b9b1fb1f28ea75 from the resource manager.
[flink-runner-job-invoker] INFO org.apache.flink.runtime.minicluster.MiniCluster - Shutting down Flink Mini Cluster
[flink-runner-job-invoker] INFO org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Shutting down rest endpoint.
[mini-cluster-io-thread-15] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - JobManager for job 3dbeef5babb92f6da7b9b1fb1f28ea75 with leader id a5396978043c970d48b812e73d0a4ceb lost leadership.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.slot.TaskSlotTable - Free slot TaskSlot(index:0, state:ACTIVE, resource profile: ResourceProfile{cpuCores=1.7976931348623157E308, heapMemoryInMB=2147483647, directMemoryInMB=2147483647, nativeMemoryInMB=2147483647, networkMemoryInMB=2147483647}, allocationId: 8348ee271798df318ba229fc43eb79d3, jobId: 3dbeef5babb92f6da7b9b1fb1f28ea75).
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.JobLeaderService - Remove job 3dbeef5babb92f6da7b9b1fb1f28ea75 from job leader monitoring.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Close JobManager connection for job 3dbeef5babb92f6da7b9b1fb1f28ea75.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Close JobManager connection for job 3dbeef5babb92f6da7b9b1fb1f28ea75.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.JobLeaderService - Cannot reconnect to job 3dbeef5babb92f6da7b9b1fb1f28ea75 because it is not registered.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Stopping TaskExecutor akka://flink/user/taskmanager_0.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.JobLeaderService - Stop job leader service.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.state.TaskExecutorLocalStateStoresManager - Shutting down TaskExecutorLocalStateStoresManager.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager removed spill file directory /tmp/flink-io-1e4ca8ab-9b77-4aa9-a891-f9db978fd094
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.io.network.NetworkEnvironment - Shutting down the network environment and its components.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.JobLeaderService - Stop job leader service.
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.filecache.FileCache - removed file cache directory /tmp/flink-dist-cache-9accd22b-d3fa-40fd-b055-de5983d260a5
[flink-akka.actor.default-dispatcher-8] INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Stopped TaskExecutor akka://flink/user/taskmanager_0.
[ForkJoinPool.commonPool-worker-2] INFO org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Removing cache directory /tmp/flink-web-ui
[ForkJoinPool.commonPool-worker-2] INFO org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Shut down complete.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.resourcemanager.StandaloneResourceManager - Shut down cluster because application is in CANCELED, diagnostics DispatcherResourceManagerComponent has been closed..
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Stopping dispatcher akka://flink/user/dispatcher.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Stopping all currently running jobs of dispatcher akka://flink/user/dispatcher.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager - Closing the SlotManager.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager - Suspending the SlotManager.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.rest.handler.legacy.backpressure.StackTraceSampleCoordinator - Shutting down stack trace sample coordinator.
[flink-akka.actor.default-dispatcher-5] INFO org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Stopped dispatcher akka://flink/user/dispatcher.
[flink-metrics-2] INFO akka.remote.RemoteActorRefProvider$RemotingTerminator - Shutting down remote daemon.
[flink-metrics-2] INFO akka.remote.RemoteActorRefProvider$RemotingTerminator - Remote daemon shut down; proceeding with flushing remote transports.
[flink-metrics-2] INFO akka.remote.RemoteActorRefProvider$RemotingTerminator - Remoting shut down.
[flink-metrics-2] INFO org.apache.flink.runtime.rpc.akka.AkkaRpcService - Stopping Akka RPC service.
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.blob.PermanentBlobCache - Shutting down BLOB cache
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.blob.TransientBlobCache - Shutting down BLOB cache
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.blob.BlobServer - Stopped BLOB server at 0.0.0.0:40567
[flink-akka.actor.default-dispatcher-4] INFO org.apache.flink.runtime.rpc.akka.AkkaRpcService - Stopped Akka RPC service.
[flink-runner-job-invoker] INFO org.apache.beam.runners.flink.FlinkPipelineRunner - Execution finished in 16361 msecs
[flink-runner-job-invoker] INFO org.apache.beam.runners.flink.FlinkPipelineRunner - Final accumulator values:
[flink-runner-job-invoker] INFO org.apache.beam.runners.flink.FlinkPipelineRunner - __metricscontainers : MetricQueryResults(Counters(ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_27:0}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_9:0}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_9:0}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_27:0}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_9:0}: 0, ref_PCollection_PCollection_17:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2474>)_26}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/ToVoidKey_30}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=external_2root/Init/Map/ParMultiDo(Anonymous)}: 0, ref_PCollection_PCollection_1:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2474>)_4}: 4, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_12:0}: 0, ref_PCollection_PCollection_12:beam:metric:element_count:v1 {PCOLLECTION=pcollection}: 5, ref_PCollection_PCollection_1:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_1:0}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ExternalTransform(beam:transforms:xlang:count)/Combine.perKey(Count)/Precombine}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/ToVoidKey_30}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:1:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_14:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:1:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_14:0}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_12:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_9:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/Map(decode)_16}: 0, ref_PCollection_PCollection_17:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_18}: 1, ref_PCollection_PCollection_17:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_19}: 1, ref_PCollection_PCollection_17:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/Map(decode)_28}: 0, ref_PCollection_PCollection_17:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_17}: 1, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ExternalTransform(beam:transforms:xlang:count)/Combine.perKey(Count)/Precombine}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=external_2root/Init/Map/ParMultiDo(Anonymous)}: 0, pcollection_1:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_14:0}: 0, pcollection_1:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_14:0}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Match_41}: 0, pcollection_1:beam:metric:element_count:v1 {PCOLLECTION=pcollection_1}: 3, ref_PCollection_PCollection_17:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_24:0}: 1, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_14:0}: 0, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_2:0}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2474>)_26}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Match_41}: 0, ref_PCollection_PCollection_1:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_1}: 1, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_29}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/ToVoidKey_30}: 0, ref_PCollection_PCollection_1:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_2}: 12, pcollection_1:beam:metric:element_count:v1 {PCOLLECTION=pcollection_2}: 3, ref_PCollection_PCollection_27:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_30}: 1, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_14:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_29}: 0, pcollection_1:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/read/pcollection_1:0}: 0, ref_PCollection_PCollection_17:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_0_32}: 11, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:378>)_22}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(unicode)_17}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ExternalTransform(beam:transforms:xlang:count)/Combine.perKey(Count)/Precombine}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:378>)_22}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Unkey_40}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:378>)_22}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:377>)_21}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=external_2root/Init/Map/ParMultiDo(Anonymous)}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_39}: 0, ref_PCollection_PCollection_9:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(unicode)_17}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_29}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2474>)_26}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:0:0}: 0, ref_PCollection_PCollection_27:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_27:0}: 0, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/ToVoidKey_30}: 0, ref_PCollection_PCollection_27:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Unkey_40}: 0, pcollection_1:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_14}: 3, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_27:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:377>)_21}: 0, pcollection_1:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ExternalTransform(beam:transforms:xlang:count)/Combine.perKey(Count)/Merge}: 0, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_1_33}: 15, ref_PCollection_PCollection_17:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_22}: 1, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2474>)_26}: 0, ref_PCollection_PCollection_27:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_28}: 1, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:1:0}: 0, ref_PCollection_PCollection_17:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_17:0}: 0, ref_PCollection_PCollection_27:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_29}: 1, ref_PCollection_PCollection_12:beam:metric:element_count:v1 {PCOLLECTION=external_2root/Init/Map/ParMultiDo(Anonymous).output}: 6, ref_PCollection_PCollection_27:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_27}: 1, pcollection_1:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/read/pcollection_1:0}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(unicode)_17}: 0, ref_PCollection_PCollection_17:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:0:0}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/Map(decode)_16}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_39}: 2, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_39}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/write/pcollection:0}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_12:0}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/write/pcollection:0}: 0, ref_PCollection_PCollection_12:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_13}: 6, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/Map(decode)_16}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/write/pcollection:0}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(unicode)_17}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Unkey_40}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_0_32}: 0, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_1:0}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_0_32}: 0, ref_PCollection_PCollection_14:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_14}: 3, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_12:0}: 0, ref_PCollection_PCollection_9:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_9}: 12, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:378>)_22}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:377>)_21}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:0:0}: 0, ref_PCollection_PCollection_14:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_16}: 3, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:0:0}: 0, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_1:0}: 0, ref_PCollection_PCollection_14:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_15}: 3, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_1:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_2:0}: 0, ref_PCollection_PCollection_17:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Flatten_34}: 0, ref_PCollection_PCollection_12:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_12}: 12, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_1:0}: 0, ref_PCollection_PCollection_9:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_9:0}: 0, pcollection_1:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ExternalTransform(beam:transforms:xlang:count)/Combine.perKey(Count)/ExtractOutputs}: 0, ref_PCollection_PCollection_14:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_24:1}: 3, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Unkey_40}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_1_33}: 0, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_24:1:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_1_33}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/Map(decode)_28}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/Map(decode)_28}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_12:0}: 0, ref_PCollection_PCollection_14:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_1_33}: 15, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_17:0}: 0, ref_PCollection_PCollection_27:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Match_41}: 0, ref_PCollection_PCollection_9:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:373>)_18}: 0, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_2:0}: 0, ref_PCollection_PCollection_9:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_12:0}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_17:0}: 0, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2474>)_4}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Create/Map(decode)_28}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:373>)_18}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/pair_with_0_32}: 11, ref_PCollection_PCollection_14:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_23}: 3, ref_PCollection_PCollection_27:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Match_41}: 0, ref_PCollection_PCollection_14:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_20}: 3, pcollection_1:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_14:0}: 0, ref_PCollection_PCollection_14:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_21}: 3, ref_PCollection_PCollection_9:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_11}: 12, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:373>)_18}: 0, ref_PCollection_PCollection_9:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_12}: 12, ref_PCollection_PCollection_9:beam:metric:element_count:v1 {PCOLLECTION=ref_PCollection_PCollection_10}: 12, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2474>)_4}: 0, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2474>)_4}: 4, ref_PCollection_PCollection_27:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_39}: 2, ref_PCollection_PCollection_1:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=fn/write/ref_PCollection_PCollection_2:0}: 0, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_29}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=external_1root/ParDo(Anonymous)/ParMultiDo(Anonymous)}: 0, ref_PCollection_PCollection_17:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=fn/read/ref_PCollection_PCollection_17:0}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:process_bundle_msecs:v1 {PTRANSFORM=external_1root/ParDo(Anonymous)/ParMultiDo(Anonymous)}: 0, ref_PCollection_PCollection_12:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=external_1root/ParDo(Anonymous)/ParMultiDo(Anonymous)}: 0, ref_PCollection_PCollection_14:beam:metric:ptransform_execution_time:total_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:377>)_21}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:start_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Create/Map(decode)_16}: 0, ref_PCollection_PCollection_9:beam:metric:pardo_execution_time:finish_bundle_msecs:v1 {PTRANSFORM=ref_AppliedPTransform_Map(<lambda at external_test.py:373>)_18}: 0)Distributions(ref_PCollection_PCollection_14:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_16}: DistributionResult{sum=54, count=3, min=18, max=18}, ref_PCollection_PCollection_14:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_15}: DistributionResult{sum=51, count=3, min=17, max=17}, ref_PCollection_PCollection_14:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_14}: DistributionResult{sum=45, count=3, min=15, max=15}, ref_PCollection_PCollection_27:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_30}: DistributionResult{sum=14, count=1, min=14, max=14}, ref_PCollection_PCollection_17:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_24:0}: DistributionResult{sum=19, count=1, min=19, max=19}, ref_PCollection_PCollection_17:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_17}: DistributionResult{sum=13, count=1, min=13, max=13}, ref_PCollection_PCollection_14:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_24:1}: DistributionResult{sum=72, count=3, min=24, max=24}, ref_PCollection_PCollection_17:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_18}: DistributionResult{sum=16, count=1, min=16, max=16}, ref_PCollection_PCollection_17:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_19}: DistributionResult{sum=15, count=1, min=15, max=15}, ref_PCollection_PCollection_9:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_9}: DistributionResult{sum=192, count=12, min=16, max=16}, ref_PCollection_PCollection_17:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_22}: DistributionResult{sum=17, count=1, min=17, max=17}, ref_PCollection_PCollection_9:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_10}: DistributionResult{sum=168, count=12, min=14, max=14}, ref_PCollection_PCollection_9:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_12}: DistributionResult{sum=168, count=12, min=14, max=14}, ref_PCollection_PCollection_9:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_11}: DistributionResult{sum=168, count=12, min=14, max=14}, ref_PCollection_PCollection_1:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_1}: DistributionResult{sum=13, count=1, min=13, max=13}, ref_PCollection_PCollection_27:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_27}: DistributionResult{sum=58, count=1, min=58, max=58}, ref_PCollection_PCollection_27:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_28}: DistributionResult{sum=41, count=1, min=41, max=41}, ref_PCollection_PCollection_14:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_23}: DistributionResult{sum=63, count=3, min=21, max=21}, ref_PCollection_PCollection_27:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_29}: DistributionResult{sum=33, count=1, min=33, max=33}, ref_PCollection_PCollection_1:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_2}: DistributionResult{sum=180, count=12, min=15, max=15}, ref_PCollection_PCollection_14:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_21}: DistributionResult{sum=57, count=3, min=19, max=19}, ref_PCollection_PCollection_14:beam:metric:sampled_byte_size:v1 {PCOLLECTION=ref_PCollection_PCollection_20}: DistributionResult{sum=54, count=3, min=18, max=18}))
[flink-runner-job-invoker] INFO org.apache.beam.runners.fnexecution.artifact.BeamFileSystemArtifactRetrievalService - Manifest at /tmp/beam-artifact-staging/job_4b84a32e-6d96-4f67-8517-4afe310ab81b/MANIFEST has 0 artifact locations
[flink-runner-job-invoker] INFO org.apache.beam.runners.fnexecution.artifact.BeamFileSystemArtifactStagingService - Removed dir /tmp/beam-artifact-staging/job_4b84a32e-6d96-4f67-8517-4afe310ab81b/
INFO:root:Job state changed to DONE

> Task :sdks:python:test-suites:portable:py2:crossLanguageTests

> Task :sdks:python:test-suites:dataflow:py2:postCommitIT
test_bigquery_tornadoes_it (apache_beam.examples.cookbook.bigquery_tornadoes_it_test.BigqueryTornadoesIT) ... ok
test_datastore_wordcount_it (apache_beam.examples.cookbook.datastore_wordcount_it_test.DatastoreWordCountIT) ... ok
test_autocomplete_it (apache_beam.examples.complete.autocomplete_test.AutocompleteTest) ... ok
test_leader_board_it (apache_beam.examples.complete.game.leader_board_it_test.LeaderBoardIT) ... ok
test_datastore_write_limit (apache_beam.io.gcp.datastore.v1new.datastore_write_it_test.DatastoreWriteIT) ... SKIP: GCP dependencies are not installed
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:695: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=transform.kms_key))
test_game_stats_it (apache_beam.examples.complete.game.game_stats_it_test.GameStatsIT) ... ok
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/bigquery_file_loads.py>:793: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  temp_location = p.options.view_as(GoogleCloudOptions).temp_location
test_wordcount_fnapi_it (apache_beam.examples.wordcount_it_test.WordCountIT) ... ok
test_wordcount_it (apache_beam.examples.wordcount_it_test.WordCountIT) ... ok
test_streaming_wordcount_it (apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT) ... ok
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/bigquery_test.py>:577: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  streaming = self.test_pipeline.options.view_as(StandardOptions).streaming
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1145: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  experiments = p.options.view_as(DebugOptions).experiments or []
test_hourly_team_score_it (apache_beam.examples.complete.game.hourly_team_score_it_test.HourlyTeamScoreIT) ... ok
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1145: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  experiments = p.options.view_as(DebugOptions).experiments or []
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/bigquery_file_loads.py>:793: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  temp_location = p.options.view_as(GoogleCloudOptions).temp_location
test_user_score_it (apache_beam.examples.complete.game.user_score_it_test.UserScoreIT) ... ok
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:1142: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported
  self.table_reference.projectId = pcoll.pipeline.options.view_as(
test_avro_it (apache_beam.examples.fastavro_it_test.FastavroIT) ... ok
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py>:695: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=transform.kms_key))
test_bigquery_read_1M_python (apache_beam.io.gcp.bigquery_io_read_it_test.BigqueryIOReadIT) ... ok
test_multiple_destinations_transform (apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests) ... ok
test_copy (apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest) ... ok
test_copy_batch (apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest) ... ok
test_copy_batch_kms (apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest) ... ok
test_copy_batch_rewrite_token (apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest) ... ok
test_copy_kms (apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest) ... ok
test_copy_rewrite_token (apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest) ... ok
test_value_provider_transform (apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests) ... ok
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/fileio_test.py>:232: FutureWarning: MatchAll is experimental.
  | 'GetPath' >> beam.Map(lambda metadata: metadata.path))
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/fileio_test.py>:243: FutureWarning: MatchAll is experimental.
  | 'Checksums' >> beam.Map(compute_hash))
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/fileio_test.py>:243: FutureWarning: ReadMatches is experimental.
  | 'Checksums' >> beam.Map(compute_hash))
test_big_query_read (apache_beam.io.gcp.bigquery_read_it_test.BigQueryReadIntegrationTests) ... ok
test_big_query_read_new_types (apache_beam.io.gcp.bigquery_read_it_test.BigQueryReadIntegrationTests) ... ok
test_bqfl_streaming (apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT) ... SKIP: TestStream is not supported on TestDataflowRunner
test_multiple_destinations_transform (apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT) ... ok
test_one_job_fails_all_jobs_fail (apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT) ... ok
test_transform_on_gcs (apache_beam.io.fileio_test.MatchIntegrationTest) ... ok
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/big_query_query_to_table_pipeline.py>:73: BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use WriteToBigQuery instead.
  kms_key=kms_key))
test_file_loads (apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT) ... SKIP: https://issuetracker.google.com/issues/118375066
test_streaming_inserts (apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT) ... ok
test_parquetio_it (apache_beam.io.parquetio_it_test.TestParquetIT) ... ok
Runs streaming Dataflow job and verifies that user metrics are reported ... ok
test_streaming_data_only (apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest) ... ok
test_streaming_with_attributes (apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest) ... ok
test_big_query_write (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests) ... ok
test_big_query_write_new_types (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests) ... ok
test_big_query_write_schema_autodetect (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests) ... SKIP: DataflowRunner does not support schema autodetection
test_big_query_write_without_schema (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests) ... ok
test_big_query_legacy_sql (apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT) ... ok
test_big_query_new_types (apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT) ... ok
test_big_query_standard_sql (apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT) ... ok
test_big_query_standard_sql_kms_key_native (apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT) ... ok
test_job_python_from_python_it (apache_beam.transforms.external_test_it.ExternalTransformIT) ... ok
test_metrics_fnapi_it (apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest) ... ok
test_metrics_it (apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest) ... ok
test_datastore_write_limit (apache_beam.io.gcp.datastore_write_it_test.DatastoreWriteIT) ... ok

----------------------------------------------------------------------
XML: nosetests-postCommitIT-df.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 45 tests in 3673.246s

OK (SKIP=4)

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/test-suites/direct/py2/build.gradle'> line: 80

* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py2:hdfsIntegrationTest'.
> Process 'command 'sh'' finished with non-zero exit value 255

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 2m 38s
111 actionable tasks: 90 executed, 18 from cache, 3 up-to-date

Publishing build scan...
https://gradle.com/s/7z4gveulgrvbm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org