You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/10/18 19:56:53 UTC

Build failed in Jenkins: beam_PostCommit_Python2 #742

See <https://builds.apache.org/job/beam_PostCommit_Python2/742/display/redirect>

Changes:


------------------------------------------
[...truncated 940.15 KB...]
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:interval_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "TeamScoresDict.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s9"
        }, 
        "serialized_fn": "<string of 492 bytes>", 
        "user_name": "TeamScoresDict"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s11", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:interval_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "WriteTeamScoreSums/ConvertToRow.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s10"
        }, 
        "serialized_fn": "<string of 1456 bytes>", 
        "user_name": "WriteTeamScoreSums/ConvertToRow"
      }
    }, 
    {
      "kind": "ParallelWrite", 
      "name": "s12", 
      "properties": {
        "create_disposition": "CREATE_IF_NEEDED", 
        "dataset": "hourly_team_score_it_dataset15714245298364", 
        "display_data": [], 
        "encoding": {
          "@type": "kind:windowed_value", 
          "component_encodings": [
            {
              "@type": "RowAsDictJsonCoder$eNprYEpOLEhMzkiNT0pNzNXLzNdLTy7QS8pMLyxNLaqML8nPzynmCsovdyx2yUwu8SrOz3POT0kt4ipk0GwsZKwtZErSAwBKpRfo", 
              "component_encodings": []
            }, 
            {
              "@type": "kind:global_window"
            }
          ], 
          "is_wrapper": true
        }, 
        "format": "bigquery", 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s11"
        }, 
        "project": "apache-beam-testing", 
        "schema": "{\"fields\": [{\"type\": \"INTEGER\", \"name\": \"total_score\", \"mode\": \"NULLABLE\"}, {\"type\": \"STRING\", \"name\": \"window_start\", \"mode\": \"NULLABLE\"}, {\"type\": \"STRING\", \"name\": \"team\", \"mode\": \"NULLABLE\"}]}", 
        "table": "leader_board", 
        "user_name": "WriteTeamScoreSums/WriteToBigQuery/WriteToBigQuery/NativeWrite", 
        "write_disposition": "WRITE_APPEND"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2019-10-18T18:49:05.180105Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2019-10-18_11_49_03-484373359420189818'
 location: u'us-central1'
 name: u'beamapp-jenkins-1018184850-350534'
 projectId: u'apache-beam-testing'
 stageStates: []
 startTime: u'2019-10-18T18:49:05.180105Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-10-18_11_49_03-484373359420189818]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-18_11_49_03-484373359420189818?project=apache-beam-testing
root: INFO: Job 2019-10-18_11_49_03-484373359420189818 is in state JOB_STATE_RUNNING
root: INFO: 2019-10-18T18:49:03.639Z: JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 2019-10-18_11_49_03-484373359420189818. The number of workers will be between 1 and 1000.
root: INFO: 2019-10-18T18:49:03.639Z: JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 2019-10-18_11_49_03-484373359420189818.
root: INFO: 2019-10-18T18:49:07.847Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-18T18:49:08.872Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-a.
root: INFO: 2019-10-18T18:49:09.585Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-10-18T18:49:09.665Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
root: INFO: 2019-10-18T18:49:09.701Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-10-18T18:49:09.772Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-10-18T18:49:09.842Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-10-18T18:49:09.871Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/ParseGameEventFn into ReadInputText/Read
root: INFO: 2019-10-18T18:49:09.906Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/FilterStartTime into HourlyTeamScore/ParseGameEventFn
root: INFO: 2019-10-18T18:49:09.942Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/FilterEndTime into HourlyTeamScore/FilterStartTime
root: INFO: 2019-10-18T18:49:09.979Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/AddEventTimestamps into HourlyTeamScore/FilterEndTime
root: INFO: 2019-10-18T18:49:10.020Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/FixedWindowsTeam into HourlyTeamScore/AddEventTimestamps
root: INFO: 2019-10-18T18:49:10.058Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/ExtractAndSumScore/Map(<lambda at hourly_team_score.py:142>) into HourlyTeamScore/FixedWindowsTeam
root: INFO: 2019-10-18T18:49:10.094Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Partial into HourlyTeamScore/ExtractAndSumScore/Map(<lambda at hourly_team_score.py:142>)
root: INFO: 2019-10-18T18:49:10.134Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Reify into HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Partial
root: INFO: 2019-10-18T18:49:10.169Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Write into HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Reify
root: INFO: 2019-10-18T18:49:10.198Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine into HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Read
root: INFO: 2019-10-18T18:49:10.233Z: JOB_MESSAGE_DETAILED: Fusing consumer HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Extract into HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine
root: INFO: 2019-10-18T18:49:10.274Z: JOB_MESSAGE_DETAILED: Fusing consumer TeamScoresDict into HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Extract
root: INFO: 2019-10-18T18:49:10.313Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteTeamScoreSums/ConvertToRow into TeamScoresDict
root: INFO: 2019-10-18T18:49:10.353Z: JOB_MESSAGE_DETAILED: Fusing consumer WriteTeamScoreSums/WriteToBigQuery/WriteToBigQuery/NativeWrite into WriteTeamScoreSums/ConvertToRow
root: INFO: 2019-10-18T18:49:10.395Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
root: INFO: 2019-10-18T18:49:10.426Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-10-18T18:49:10.463Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-10-18T18:49:10.502Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-10-18T18:49:10.668Z: JOB_MESSAGE_DEBUG: Executing wait step start22
root: INFO: 2019-10-18T18:49:10.754Z: JOB_MESSAGE_BASIC: Executing operation HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Create
root: INFO: 2019-10-18T18:49:10.798Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-10-18T18:49:10.832Z: JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-a...
root: INFO: 2019-10-18T18:49:10.900Z: JOB_MESSAGE_BASIC: Finished operation HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Create
root: INFO: 2019-10-18T18:49:11.192Z: JOB_MESSAGE_DEBUG: Value "HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Session" materialized.
root: INFO: 2019-10-18T18:49:11.304Z: JOB_MESSAGE_BASIC: Executing operation ReadInputText/Read+HourlyTeamScore/ParseGameEventFn+HourlyTeamScore/FilterStartTime+HourlyTeamScore/FilterEndTime+HourlyTeamScore/AddEventTimestamps+HourlyTeamScore/FixedWindowsTeam+HourlyTeamScore/ExtractAndSumScore/Map(<lambda at hourly_team_score.py:142>)+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Partial+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Reify+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Write
root: INFO: 2019-10-18T18:49:38.163Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 1 to 5.
root: INFO: 2019-10-18T18:49:42.481Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 1 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T18:49:42.519Z: JOB_MESSAGE_DETAILED: Resized worker pool to 1, though goal was 5.  This could be a quota issue.
root: INFO: 2019-10-18T18:50:09.099Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 5 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T18:50:53.316Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
root: INFO: 2019-10-18T18:50:53.359Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
root: INFO: 2019-10-18T18:54:19.040Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 5 to 78.
root: INFO: 2019-10-18T18:54:24.860Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 26 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T18:54:24.889Z: JOB_MESSAGE_DETAILED: Resized worker pool to 26, though goal was 78.  This could be a quota issue.
root: INFO: 2019-10-18T18:54:30.235Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 71 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T18:54:30.275Z: JOB_MESSAGE_DETAILED: Resized worker pool to 71, though goal was 78.  This could be a quota issue.
root: INFO: 2019-10-18T18:54:35.616Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 78 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T18:55:10.620Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-18T18:57:48.073Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 46 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T18:57:53.116Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 78 to 46.
root: INFO: 2019-10-18T18:58:49.723Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 46 to 57.
root: INFO: 2019-10-18T18:58:55.339Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 55 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T18:58:55.362Z: JOB_MESSAGE_DETAILED: Resized worker pool to 55, though goal was 57.  This could be a quota issue.
root: INFO: 2019-10-18T18:59:00.665Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 57 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T19:01:10.620Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-18T19:01:49.700Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 57 to 63.
root: INFO: 2019-10-18T19:02:00.715Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 58 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T19:02:00.755Z: JOB_MESSAGE_DETAILED: Resized worker pool to 58, though goal was 63.  This could be a quota issue.
root: INFO: 2019-10-18T19:02:06.171Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 62 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T19:02:06.202Z: JOB_MESSAGE_DETAILED: Resized worker pool to 62, though goal was 63.  This could be a quota issue.
root: INFO: 2019-10-18T19:02:16.922Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 63 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T19:03:51.919Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 63 to 75.
root: INFO: 2019-10-18T19:03:57.627Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 66 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T19:03:57.658Z: JOB_MESSAGE_DETAILED: Resized worker pool to 66, though goal was 75.  This could be a quota issue.
root: INFO: 2019-10-18T19:04:03.020Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 75 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T19:04:21.984Z: JOB_MESSAGE_ERROR: Workflow failed.
root: INFO: 2019-10-18T19:04:22.078Z: JOB_MESSAGE_BASIC: Finished operation ReadInputText/Read+HourlyTeamScore/ParseGameEventFn+HourlyTeamScore/FilterStartTime+HourlyTeamScore/FilterEndTime+HourlyTeamScore/AddEventTimestamps+HourlyTeamScore/FixedWindowsTeam+HourlyTeamScore/ExtractAndSumScore/Map(<lambda at hourly_team_score.py:142>)+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/Combine/Partial+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Reify+HourlyTeamScore/ExtractAndSumScore/CombinePerKey(sum)/GroupByKey/Write
root: INFO: 2019-10-18T19:04:22.703Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2019-10-18T19:04:22.769Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown.
root: INFO: 2019-10-18T19:04:22.804Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2019-10-18T19:08:18.406Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-18T19:08:18.455Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2019-10-18T19:08:18.493Z: JOB_MESSAGE_DEBUG: Tearing down pending resources...
root: INFO: Job 2019-10-18_11_49_03-484373359420189818 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: nosetests-postCommitIT-df.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 45 tests in 4083.873s

FAILED (SKIP=4, errors=1)

> Task :sdks:python:test-suites:dataflow:py2:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/test-suites/dataflow/py2/build.gradle'> line: 85

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py2:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 9m 0s
111 actionable tasks: 86 executed, 22 from cache, 3 up-to-date

Publishing build scan...
https://gradle.com/s/3xowaiun3naug

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Python2 #744

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python2/744/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PostCommit_Python2 #743

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python2/743/display/redirect?page=changes>

Changes:

[github] Adds a section on how to make reviewer's job easier in Contributor Guide


------------------------------------------
[...truncated 940.05 KB...]
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s16"
        }, 
        "serialized_fn": "<string of 1368 bytes>", 
        "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key)"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s18", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Unkey.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s17"
        }, 
        "serialized_fn": "<string of 960 bytes>", 
        "user_name": "assert_that/Unkey"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s19", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s18"
        }, 
        "serialized_fn": "<string of 1336 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2019-10-18T22:04:15.028573Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2019-10-18_15_04_13-12869122748529240285'
 location: u'us-central1'
 name: u'beamapp-jenkins-1018220404-540935'
 projectId: u'apache-beam-testing'
 stageStates: []
 startTime: u'2019-10-18T22:04:15.028573Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-10-18_15_04_13-12869122748529240285]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-18_15_04_13-12869122748529240285?project=apache-beam-testing
root: INFO: Job 2019-10-18_15_04_13-12869122748529240285 is in state JOB_STATE_RUNNING
root: INFO: 2019-10-18T22:04:13.278Z: JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 2019-10-18_15_04_13-12869122748529240285.
root: INFO: 2019-10-18T22:04:13.278Z: JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 2019-10-18_15_04_13-12869122748529240285. The number of workers will be between 1 and 1000.
root: INFO: 2019-10-18T22:04:17.113Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-18T22:04:17.922Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-a.
root: INFO: 2019-10-18T22:04:18.584Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-10-18T22:04:18.617Z: JOB_MESSAGE_DEBUG: Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2019-10-18T22:04:18.703Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
root: INFO: 2019-10-18T22:04:18.742Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-10-18T22:04:18.848Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-10-18T22:04:18.986Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-10-18T22:04:19.026Z: JOB_MESSAGE_DETAILED: Fusing consumer row to string into read
root: INFO: 2019-10-18T22:04:19.065Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/KeyWithVoid into row to string
root: INFO: 2019-10-18T22:04:19.103Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey+count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Partial into count/CombineGlobally(CountCombineFn)/KeyWithVoid
root: INFO: 2019-10-18T22:04:19.140Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Reify into count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey+count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Partial
root: INFO: 2019-10-18T22:04:19.176Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Write into count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Reify
root: INFO: 2019-10-18T22:04:19.211Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine into count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Read
root: INFO: 2019-10-18T22:04:19.247Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Extract into count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine
root: INFO: 2019-10-18T22:04:19.285Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/UnKey into count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Extract
root: INFO: 2019-10-18T22:04:19.318Z: JOB_MESSAGE_DETAILED: Unzipping flatten s15 for input s13.out
root: INFO: 2019-10-18T22:04:19.353Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten assert_that/Group/Flatten, into producer assert_that/Group/pair_with_0
root: INFO: 2019-10-18T22:04:19.390Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read
root: INFO: 2019-10-18T22:04:19.425Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2019-10-18T22:04:19.458Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2019-10-18T22:04:19.496Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Match into assert_that/Unkey
root: INFO: 2019-10-18T22:04:19.535Z: JOB_MESSAGE_DETAILED: Unzipping flatten s15-u31 for input s16-reify-value9-c29
root: INFO: 2019-10-18T22:04:19.574Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten assert_that/Group/Flatten/Unzipped-1, into producer assert_that/Group/GroupByKey/Reify
root: INFO: 2019-10-18T22:04:19.595Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2019-10-18T22:04:19.633Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1
root: INFO: 2019-10-18T22:04:19.665Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2019-10-18T22:04:19.703Z: JOB_MESSAGE_DETAILED: Fusing consumer count/CombineGlobally(CountCombineFn)/InjectDefault/InjectDefault into count/CombineGlobally(CountCombineFn)/DoOnce/Read
root: INFO: 2019-10-18T22:04:19.741Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/WindowInto(WindowIntoFn) into count/CombineGlobally(CountCombineFn)/InjectDefault/InjectDefault
root: INFO: 2019-10-18T22:04:19.774Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2019-10-18T22:04:19.814Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2019-10-18T22:04:19.854Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
root: INFO: 2019-10-18T22:04:19.893Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-10-18T22:04:19.928Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-10-18T22:04:19.966Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-10-18T22:04:20.135Z: JOB_MESSAGE_DEBUG: Executing wait step start41
root: INFO: 2019-10-18T22:04:20.224Z: JOB_MESSAGE_BASIC: Executing operation assert_that/Group/GroupByKey/Create
root: INFO: 2019-10-18T22:04:20.259Z: JOB_MESSAGE_BASIC: Executing operation count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Create
root: INFO: 2019-10-18T22:04:20.272Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-10-18T22:04:20.305Z: JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-a...
root: INFO: 2019-10-18T22:04:20.347Z: JOB_MESSAGE_BASIC: Finished operation assert_that/Group/GroupByKey/Create
root: INFO: 2019-10-18T22:04:20.361Z: JOB_MESSAGE_BASIC: Finished operation count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Create
root: INFO: 2019-10-18T22:04:20.423Z: JOB_MESSAGE_DEBUG: Value "assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2019-10-18T22:04:20.462Z: JOB_MESSAGE_DEBUG: Value "count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Session" materialized.
root: INFO: 2019-10-18T22:04:20.510Z: JOB_MESSAGE_BASIC: Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2019-10-18T22:04:20.554Z: JOB_MESSAGE_BASIC: Executing operation read+row to string+count/CombineGlobally(CountCombineFn)/KeyWithVoid+count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey+count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Partial+count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Reify+count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Write
root: INFO: 2019-10-18T22:04:21.081Z: JOB_MESSAGE_BASIC: BigQuery export job "dataflow_job_6435351174498814187" started. You can check its status with the bq tool: "bq show -j --project_id=apache-beam-testing dataflow_job_6435351174498814187".
root: INFO: 2019-10-18T22:04:51.432Z: JOB_MESSAGE_DETAILED: BigQuery export job progress: "dataflow_job_6435351174498814187" observed total of 1 exported files thus far.
root: INFO: 2019-10-18T22:04:51.470Z: JOB_MESSAGE_BASIC: BigQuery export job finished: "dataflow_job_6435351174498814187"
root: INFO: 2019-10-18T22:10:20.108Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-18T22:16:20.107Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-18T22:17:06.526Z: JOB_MESSAGE_ERROR: Startup of the worker pool in zone us-central1-a failed to bring up any of the desired 1 workers. INTERNAL_ERROR: Internal error. Please try again or contact Google Support. (Code: '-5977583105891360131')
root: INFO: 2019-10-18T22:17:06.545Z: JOB_MESSAGE_ERROR: Workflow failed.
root: INFO: 2019-10-18T22:17:06.603Z: JOB_MESSAGE_BASIC: Finished operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2019-10-18T22:17:06.604Z: JOB_MESSAGE_BASIC: Finished operation read+row to string+count/CombineGlobally(CountCombineFn)/KeyWithVoid+count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey+count/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Partial+count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Reify+count/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Write
root: INFO: 2019-10-18T22:17:06.742Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2019-10-18T22:17:06.841Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown.
root: INFO: 2019-10-18T22:17:06.879Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2019-10-18T22:17:25.548Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2019-10-18T22:17:25.578Z: JOB_MESSAGE_DEBUG: Tearing down pending resources...
root: INFO: Job 2019-10-18_15_04_13-12869122748529240285 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: nosetests-postCommitIT-df.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 45 tests in 4097.483s

FAILED (SKIP=4, errors=1)

> Task :sdks:python:test-suites:dataflow:py2:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/test-suites/dataflow/py2/build.gradle'> line: 85

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py2:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 9m 10s
111 actionable tasks: 86 executed, 22 from cache, 3 up-to-date

Publishing build scan...
https://gradle.com/s/lifo5i7mxhmie

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org