You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/03/26 18:37:29 UTC
Build failed in Jenkins: beam_PostCommit_XVR_Spark #455
See <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/455/display/redirect?page=changes>
Changes:
[robertwb] [BEAM-9340] Validate pipeline requirements in PipelineValidator.
[mxm] [BEAM-9566] Mitigate performance issue for output timestamp watermark
------------------------------------------
[...truncated 641.57 KB...]
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_combine_per_key (apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.21.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7ff92516b230> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:24 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_Create/Impulse_3\n Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2644>)_4\n Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/AddRandomKeys_7\n Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_9\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_10\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys_15\n Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/Map(decode)_16\n Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/Precombine\n ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/Precombine:beam:transform:combine_per_key_precombine:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/Group\n ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/Group:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/Merge\n ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/Merge:beam:transform:combine_per_key_merge_accumulators:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/ExtractOutputs\n ExternalTransform(beam:transforms:xlang:test:compk)/CombinePerKey(sum)/ExtractOutputs:beam:transform:combine_per_key_extract_outputs:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_20\n assert_that/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2644>)_21\n assert_that/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_23\n assert_that/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_24\n assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_25\n assert_that/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_27\n assert_that/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_28\n assert_that/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_29\n assert_that/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_30\n assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_34\n assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_35\n assert_that/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_36\n assert_that/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_flatten (apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.21.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7ff92516b230> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:29 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_col1/Impulse_3\n col1/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col1/FlatMap(<lambda at core.py:2644>)_4\n col1/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col1/MaybeReshuffle/Reshuffle/AddRandomKeys_7\n col1/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col1/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_9\n col1/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col1/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_10\n col1/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col1/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n col1/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col1/MaybeReshuffle/Reshuffle/RemoveRandomKeys_15\n col1/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col1/Map(decode)_16\n col1/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/Impulse_18\n col2/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/FlatMap(<lambda at core.py:2644>)_19\n col2/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/MaybeReshuffle/Reshuffle/AddRandomKeys_22\n col2/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_24\n col2/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_25\n col2/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_29\n col2/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/MaybeReshuffle/Reshuffle/RemoveRandomKeys_30\n col2/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_col2/Map(decode)_31\n col2/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_6_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:flatten)/Flatten_3\n ExternalTransform(beam:transforms:xlang:test:flatten)/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_35\n assert_that/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2644>)_36\n assert_that/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_38\n assert_that/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_39\n assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_40\n assert_that/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_42\n assert_that/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_43\n assert_that/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_44\n assert_that/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_45\n assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_49\n assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_50\n assert_that/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_51\n assert_that/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_group_by_key (apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.21.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7ff92516b230> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:17 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_Create/Impulse_3\n Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2644>)_4\n Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/Map(decode)_6\n Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_7_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:gbk)/TestLabel_3\n ExternalTransform(beam:transforms:xlang:test:gbk)/TestLabel:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at validate_runner_xlang_test.py:103>)_8\n Map(<lambda at validate_runner_xlang_test.py:103>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_11\n assert_that/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2644>)_12\n assert_that/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_14\n assert_that/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_15\n assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_16\n assert_that/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_18\n assert_that/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_19\n assert_that/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_20\n assert_that/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_21\n assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_25\n assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_26\n assert_that/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_27\n assert_that/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_multi_input_output_with_sideinput (apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.21.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7ff92516b230> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:36 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_Main1/Impulse_3\n Main1/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Main1/FlatMap(<lambda at core.py:2644>)_4\n Main1/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Main1/Map(decode)_6\n Main1/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Main2/Impulse_8\n Main2/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Main2/FlatMap(<lambda at core.py:2644>)_9\n Main2/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Main2/Map(decode)_11\n Main2/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Side/Impulse_13\n Side/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Side/FlatMap(<lambda at core.py:2644>)_14\n Side/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Side/Map(decode)_16\n Side/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_8_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:multi)/Flatten_3\n ExternalTransform(beam:transforms:xlang:test:multi)/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_8_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:multi)/Map(<lambda at expansion_service_test.py:115>)_4\n ExternalTransform(beam:transforms:xlang:test:multi)/Map(<lambda at expansion_service_test.py:115>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_8_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:multi)/Map(<lambda at expansion_service_test.py:118>)_5\n ExternalTransform(beam:transforms:xlang:test:multi)/Map(<lambda at expansion_service_test.py:118>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_20\n assert_that/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2644>)_21\n assert_that/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_23\n assert_that/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_24\n assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_25\n assert_that/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_27\n assert_that/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_28\n assert_that/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_29\n assert_that/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_30\n assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_34\n assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_35\n assert_that/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_36\n assert_that/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Create/Impulse_39\n CheckSide/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Create/FlatMap(<lambda at core.py:2644>)_40\n CheckSide/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Create/Map(decode)_42\n CheckSide/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/WindowInto(WindowIntoFn)_43\n CheckSide/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/ToVoidKey_44\n CheckSide/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Group/pair_with_0_46\n CheckSide/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Group/pair_with_1_47\n CheckSide/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Group/Flatten_48\n CheckSide/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Group/GroupByKey_49\n CheckSide/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Group/Map(_merge_tagged_vals_under_key)_53\n CheckSide/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Unkey_54\n CheckSide/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CheckSide/Match_55\n CheckSide/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_partition (apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.21.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7ff92516b230> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:35 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_Create/Impulse_3\n Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2644>)_4\n Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/AddRandomKeys_7\n Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_9\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_10\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys_15\n Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/Map(decode)_16\n Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_9_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:partition)/Partition(CallableWrapperPartitionFn)/ParDo(ApplyPartitionFnFn)/ParDo(ApplyPartitionFnFn)_5\n ExternalTransform(beam:transforms:xlang:test:partition)/Partition(CallableWrapperPartitionFn)/ParDo(ApplyPartitionFnFn)/ParDo(ApplyPartitionFnFn):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_9_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:partition)/Map(<lambda at expansion_service_test.py:216>)_6\n ExternalTransform(beam:transforms:xlang:test:partition)/Map(<lambda at expansion_service_test.py:216>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_9_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:partition)/Map(<lambda at expansion_service_test.py:217>)_7\n ExternalTransform(beam:transforms:xlang:test:partition)/Map(<lambda at expansion_service_test.py:217>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Create/Impulse_20\n check_even/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Create/FlatMap(<lambda at core.py:2644>)_21\n check_even/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Create/Map(decode)_23\n check_even/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/WindowInto(WindowIntoFn)_24\n check_even/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/ToVoidKey_25\n check_even/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Group/pair_with_0_27\n check_even/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Group/pair_with_1_28\n check_even/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Group/Flatten_29\n check_even/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Group/GroupByKey_30\n check_even/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Group/Map(_merge_tagged_vals_under_key)_34\n check_even/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Unkey_35\n check_even/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_even/Match_36\n check_even/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Create/Impulse_39\n check_odd/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Create/FlatMap(<lambda at core.py:2644>)_40\n check_odd/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Create/Map(decode)_42\n check_odd/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/WindowInto(WindowIntoFn)_43\n check_odd/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/ToVoidKey_44\n check_odd/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Group/pair_with_0_46\n check_odd/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Group/pair_with_1_47\n check_odd/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Group/Flatten_48\n check_odd/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Group/GroupByKey_49\n check_odd/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Group/Map(_merge_tagged_vals_under_key)_53\n check_odd/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Unkey_54\n check_odd/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_check_odd/Match_55\n check_odd/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
test_prefix (apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... WARNING:root:Make sure that locally built Python SDK docker image has Python 2.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python2.7_sdk:2.21.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7ff92516b230> ====================
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:21 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
DEBUG:apache_beam.runners.portability.fn_api_runner.translations:Stages: ['ref_AppliedPTransform_Create/Impulse_3\n Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/FlatMap(<lambda at core.py:2644>)_4\n Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/AddRandomKeys_7\n Create/MaybeReshuffle/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_9\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey_10\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n Create/MaybeReshuffle/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys_15\n Create/MaybeReshuffle/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Create/Map(decode)_16\n Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'external_10_AppliedPTransform_ExternalTransform(beam:transforms:xlang:test:prefix)/TestLabel_3\n ExternalTransform(beam:transforms:xlang:test:prefix)/TestLabel:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_20\n assert_that/Create/Impulse:beam:transform:impulse:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2644>)_21\n assert_that/Create/FlatMap(<lambda at core.py:2644>):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_23\n assert_that/Create/Map(decode):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_24\n assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_25\n assert_that/ToVoidKey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_27\n assert_that/Group/pair_with_0:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_28\n assert_that/Group/pair_with_1:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_29\n assert_that/Group/Flatten:beam:transform:flatten:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_30\n assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_34\n assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_35\n assert_that/Unkey:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_36\n assert_that/Match:beam:transform:pardo:v1\n must follow: \n downstream_side_inputs: <unknown>']
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_name' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'runner' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'temp_location' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'experiments' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'streaming' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'dataflow_kms_key' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'enable_streaming_engine' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'project' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_region' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'worker_zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'zone' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'environment_cache_millis' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'files_to_stage' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'job_endpoint' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'output_executable_path' was already added
DEBUG:apache_beam.runners.portability.portable_runner:Runner option 'sdk_worker_parallelism' was already added
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
ok
----------------------------------------------------------------------
XML: nosetests-xlangValidateRunner.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 10 tests in 108.369s
OK
> Task :runners:spark:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:spark:job-server:sparkJobServerCleanup
> Task :runners:spark:job-server:validatesCrossLanguageRunner
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD SUCCESSFUL in 22m 38s
104 actionable tasks: 79 executed, 23 from cache, 2 up-to-date
Publishing build scan...
https://gradle.com/s/fhdhfoowmcfqy
Build step 'Invoke Gradle script' changed build result to SUCCESS
[beam_PostCommit_XVR_Spark] $ /bin/bash -xe /tmp/jenkins1167439855633193545.sh
+ echo src RUN CROSS-LANGUAGE SPARK USING PYTHON 3.5 src
src RUN CROSS-LANGUAGE SPARK USING PYTHON 3.5 src
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -PpythonVersion=3.5 :runners:spark:job-server:validatesCrossLanguageRunner
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy UP-TO-DATE
> Task :buildSrc:pluginDescriptors UP-TO-DATE
> Task :buildSrc:processResources UP-TO-DATE
> Task :buildSrc:classes UP-TO-DATE
> Task :buildSrc:jar UP-TO-DATE
> Task :buildSrc:assemble UP-TO-DATE
> Task :buildSrc:spotlessGroovy UP-TO-DATE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata UP-TO-DATE
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties UP-TO-DATE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build UP-TO-DATE
Configuration on demand is an incubating feature.
FAILURE: Build failed with an exception.
* What went wrong:
Could not determine the dependencies of task ':sdks:java:testing:expansion-service:buildTestExpansionServiceJar'.
> Could not resolve all dependencies for configuration ':sdks:java:testing:expansion-service:testRuntime'.
> Could not resolve net.minidev:json-smart:[1.3.1,2.3].
Required by:
project :sdks:java:testing:expansion-service > org.apache.hadoop:hadoop-client:2.8.5 > org.apache.hadoop:hadoop-common:2.8.5 > org.apache.hadoop:hadoop-auth:2.8.5 > com.nimbusds:nimbus-jose-jwt:4.41.1
> Could not resolve net.minidev:json-smart:2.3-SNAPSHOT.
> Unable to load Maven meta-data from https://oss.sonatype.org/content/repositories/staging/net/minidev/json-smart/2.3-SNAPSHOT/maven-metadata.xml.
> Could not get resource 'https://oss.sonatype.org/content/repositories/staging/net/minidev/json-smart/2.3-SNAPSHOT/maven-metadata.xml'.
> Could not GET 'https://oss.sonatype.org/content/repositories/staging/net/minidev/json-smart/2.3-SNAPSHOT/maven-metadata.xml'. Received status code 502 from server: Bad Gateway
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 5s
Publishing build scan...
https://gradle.com/s/tucnvhf57bzi6
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Jenkins build is back to normal : beam_PostCommit_XVR_Spark #457
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/457/display/redirect?page=changes>
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins: beam_PostCommit_XVR_Spark #456
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/456/display/redirect?page=changes>
Changes:
[robertwb] Add base SDK version to environment capabilities for Python and Java.
[robertwb] [BEAM-9614] Add SDK id for go.
[github] [BEAM-9495] Make DataCatalogTableProvider AutoCloseable (#11116)
------------------------------------------
Started by GitHub push by robertwb
Started by GitHub push by robertwb
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-14 (beam) in workspace <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/>
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
> git init <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
> git --version # timeout=10
> git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/*
> git config remote.origin.url https://github.com/apache/beam.git # timeout=10
> git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
> git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
> git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
> git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 0ea3ec261a4d2813e763058861a6ea723fd7f533 (origin/master)
> git config core.sparsecheckout # timeout=10
> git checkout -f 0ea3ec261a4d2813e763058861a6ea723fd7f533
Commit message: "[BEAM-9495] Make DataCatalogTableProvider AutoCloseable (#11116)"
> git rev-list --no-walk 589a3037f08920601dc53a7ced9e175142f03b5c # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content
SPARK_LOCAL_IP=127.0.0.1
[EnvInject] - Variables injected successfully.
[beam_PostCommit_XVR_Spark] $ /bin/bash -xe /tmp/jenkins8623983056043215843.sh
+ echo src RUN CROSS-LANGUAGE SPARK USING PYTHON 2.7 src
src RUN CROSS-LANGUAGE SPARK USING PYTHON 2.7 src
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/gradlew> --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -PpythonVersion=2.7 :runners:spark:job-server:validatesCrossLanguageRunner
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.
FAILURE: Build failed with an exception.
* What went wrong:
Could not determine the dependencies of task ':runners:spark:job-server:shadowJar'.
> Could not resolve all dependencies for configuration ':runners:spark:job-server:runtimeClasspath'.
> Could not resolve net.minidev:json-smart:[1.3.1,2.3].
Required by:
project :runners:spark:job-server > project :runners:spark > org.apache.hadoop:hadoop-common:2.8.5 > org.apache.hadoop:hadoop-auth:2.8.5 > com.nimbusds:nimbus-jose-jwt:4.41.1
> Could not resolve net.minidev:json-smart:2.3-SNAPSHOT.
> Unable to load Maven meta-data from https://oss.sonatype.org/content/repositories/staging/net/minidev/json-smart/2.3-SNAPSHOT/maven-metadata.xml.
> Could not get resource 'https://oss.sonatype.org/content/repositories/staging/net/minidev/json-smart/2.3-SNAPSHOT/maven-metadata.xml'.
> Could not GET 'https://oss.sonatype.org/content/repositories/staging/net/minidev/json-smart/2.3-SNAPSHOT/maven-metadata.xml'. Received status code 502 from server: Bad Gateway
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 21s
Publishing build scan...
https://gradle.com/s/h6hbgcim7vfcw
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org