You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/06/12 16:27:18 UTC

Build failed in Jenkins: beam_LoadTests_Python_Combine_Dataflow_Batch #1

See <https://builds.apache.org/job/beam_LoadTests_Python_Combine_Dataflow_Batch/1/display/redirect>

------------------------------------------
[...truncated 151.43 KB...]
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "Combine with Top 0/InjectDefault.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s7"
        }, 
        "serialized_fn": "<string of 1824 bytes>", 
        "user_name": "Combine with Top 0/InjectDefault/InjectDefault"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s10", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "_GetElement", 
            "type": "STRING", 
            "value": "apache_beam.testing.load_tests.combine_test._GetElement"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "Consume 0.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s9"
        }, 
        "serialized_fn": "<string of 432 bytes>", 
        "user_name": "Consume 0"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s11", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "MeasureTime", 
            "type": "STRING", 
            "value": "apache_beam.testing.load_tests.load_test_metrics_utils.MeasureTime"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "Measure time: End 0.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s10"
        }, 
        "serialized_fn": "<string of 504 bytes>", 
        "user_name": "Measure time: End 0"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2019-06-12T16:08:10.519660Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2019-06-12_09_08_09-7101016922272079614'
 location: u'us-central1'
 name: u'load-tests-python-dataflow-batch-combine-3-0612140441'
 projectId: u'apache-beam-testing'
 stageStates: []
 startTime: u'2019-06-12T16:08:10.519660Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-06-12_09_08_09-7101016922272079614]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-06-12_09_08_09-7101016922272079614?project=apache-beam-testing
root: INFO: Job 2019-06-12_09_08_09-7101016922272079614 is in state JOB_STATE_PENDING
root: INFO: 2019-06-12T16:08:09.600Z: JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 2019-06-12_09_08_09-7101016922272079614. The number of workers will be between 1 and 5.
root: INFO: 2019-06-12T16:08:09.653Z: JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 2019-06-12_09_08_09-7101016922272079614.
root: INFO: 2019-06-12T16:08:12.403Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-06-12T16:08:12.798Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-b.
root: INFO: 2019-06-12T16:08:13.400Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-06-12T16:08:13.456Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
root: INFO: 2019-06-12T16:08:13.500Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-06-12T16:08:13.593Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-06-12T16:08:13.653Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-06-12T16:08:13.687Z: JOB_MESSAGE_DETAILED: Fusing consumer Measure time: Start into Read
root: INFO: 2019-06-12T16:08:13.736Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/GroupByKey+Combine with Top 0/CombinePerKey/Combine/Partial into Combine with Top 0/KeyWithVoid
root: INFO: 2019-06-12T16:08:13.774Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/Combine/Extract into Combine with Top 0/CombinePerKey/Combine
root: INFO: 2019-06-12T16:08:13.822Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/Combine into Combine with Top 0/CombinePerKey/GroupByKey/Read
root: INFO: 2019-06-12T16:08:13.859Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/KeyWithVoid into Measure time: Start
root: INFO: 2019-06-12T16:08:13.898Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/GroupByKey/Write into Combine with Top 0/CombinePerKey/GroupByKey/Reify
root: INFO: 2019-06-12T16:08:13.945Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/GroupByKey/Reify into Combine with Top 0/CombinePerKey/GroupByKey+Combine with Top 0/CombinePerKey/Combine/Partial
root: INFO: 2019-06-12T16:08:13.994Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/UnKey into Combine with Top 0/CombinePerKey/Combine/Extract
root: INFO: 2019-06-12T16:08:14.033Z: JOB_MESSAGE_DETAILED: Fusing consumer Measure time: End 0 into Consume 0
root: INFO: 2019-06-12T16:08:14.068Z: JOB_MESSAGE_DETAILED: Fusing consumer Consume 0 into Combine with Top 0/InjectDefault/InjectDefault
root: INFO: 2019-06-12T16:08:14.120Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/InjectDefault/InjectDefault into Combine with Top 0/DoOnce/Read
root: INFO: 2019-06-12T16:08:14.172Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
root: INFO: 2019-06-12T16:08:14.211Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-06-12T16:08:14.257Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-06-12T16:08:14.317Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-06-12T16:08:14.511Z: JOB_MESSAGE_DEBUG: Executing wait step start23
root: INFO: 2019-06-12T16:08:14.602Z: JOB_MESSAGE_BASIC: Executing operation Combine with Top 0/CombinePerKey/GroupByKey/Create
root: INFO: 2019-06-12T16:08:14.671Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-06-12T16:08:14.720Z: JOB_MESSAGE_BASIC: Starting 5 workers in us-central1-b...
root: INFO: 2019-06-12T16:08:14.808Z: JOB_MESSAGE_DEBUG: Value "Combine with Top 0/CombinePerKey/GroupByKey/Session" materialized.
root: INFO: 2019-06-12T16:08:14.892Z: JOB_MESSAGE_BASIC: Executing operation Read+Measure time: Start+Combine with Top 0/KeyWithVoid+Combine with Top 0/CombinePerKey/GroupByKey+Combine with Top 0/CombinePerKey/Combine/Partial+Combine with Top 0/CombinePerKey/GroupByKey/Reify+Combine with Top 0/CombinePerKey/GroupByKey/Write
root: INFO: Job 2019-06-12_09_08_09-7101016922272079614 is in state JOB_STATE_RUNNING
root: INFO: 2019-06-12T16:09:18.592Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 5 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-12T16:09:43.909Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
root: INFO: 2019-06-12T16:09:43.945Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
root: INFO: 2019-06-12T16:11:44.530Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 1 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-12T16:11:49.579Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 5 to 1.
root: INFO: 2019-06-12T16:21:21.573Z: JOB_MESSAGE_BASIC: Autoscaling: Resizing worker pool from 1 to 2.
root: INFO: 2019-06-12T16:21:49.030Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 2 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-12T16:24:17.059Z: JOB_MESSAGE_DEBUG: Executing failure step failure22
root: INFO: 2019-06-12T16:24:17.107Z: JOB_MESSAGE_ERROR: Workflow failed. Causes: S02:Read+Measure time: Start+Combine with Top 0/KeyWithVoid+Combine with Top 0/CombinePerKey/GroupByKey+Combine with Top 0/CombinePerKey/Combine/Partial+Combine with Top 0/CombinePerKey/GroupByKey/Reify+Combine with Top 0/CombinePerKey/GroupByKey/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers: 
  load-tests-python-dataflo-06120908-ngzm-harness-40jg,
  load-tests-python-dataflo-06120908-ngzm-harness-40jg,
  load-tests-python-dataflo-06120908-ngzm-harness-40jg,
  load-tests-python-dataflo-06120908-ngzm-harness-40jg
root: INFO: 2019-06-12T16:24:17.298Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2019-06-12T16:24:17.362Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown.
root: INFO: 2019-06-12T16:24:17.416Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2019-06-12T16:27:07.444Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on the rate of progress in the currently running step(s).
root: INFO: 2019-06-12T16:27:07.502Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2019-06-12T16:27:07.543Z: JOB_MESSAGE_DEBUG: Tearing down pending resources...
root: INFO: Job 2019-06-12_09_08_09-7101016922272079614 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_LoadTests_Python_Combine_Dataflow_Batch/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 1148.232s

FAILED (errors=1)

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_Combine_Dataflow_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 49

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 19m 19s
4 actionable tasks: 3 executed, 1 up-to-date

Publishing build scan...
https://gradle.com/s/nqvpdbara63u4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_LoadTests_Python_Combine_Dataflow_Batch #3

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_Combine_Dataflow_Batch/3/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_Combine_Dataflow_Batch #2

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_Combine_Dataflow_Batch/2/display/redirect?page=changes>

Changes:

[lukasz.gajowy] [BEAM-7368] Add method to prepare SDK harness image for jenkins jobs

[lukasz.gajowy] [BEAM-7368] Add method to prepare Flink Job server's image for jenkins

[lukasz.gajowy] [BEAM-7368] Add cluster Flink creation and teardown steps

[lukasz.gajowy] [BEAM-7368] Add GBK load test for Python + Flink

[lukasz.gajowy] [BEAM-7368] Specify a cron job for the load test

[lukasz.gajowy] [BEAM-7368] Apply code review suggestions

[lukasz.gajowy] [BEAM-7368] Use post build script in Jenkins

[hsuryawirawan] Add side input and side output java katas

[hsuryawirawan] Add side input and side output python katas

[hsuryawirawan] Modify the TestPipeline declaration to be 'final transient'

[hsuryawirawan] Add Create.of transform names in CoGroupByKey kata

[hsuryawirawan] Improve the task description of ParDo OneToMany to describe

[ehudm] [BEAM-7432] Set input and output files for sdist

[hsuryawirawan] Update offset for some tasks

[lukasz.gajowy] [BEAM-7368] Allow defining number of slots per worker. Set to 1 by

[lukasz.gajowy] [BEAM-7368] Added parameters that allow using different flink versions

[lukasz.gajowy] [BEAM-7368] increase worker count to be equal to Flink's parallelism

[pekopeko] add compatibility check badges to README

[hsuryawirawan] Update the task descriptions: formatting, conventions, and styles

[hsuryawirawan] Update Beam version to v2.13.0

[hsuryawirawan] Add BinaryCombineFn Lambda java kata

[hsuryawirawan] Add packages for all Java katas

[hsuryawirawan] Add composite transform Java kata

[hsuryawirawan] Add composite transform Python kata

[hsuryawirawan] Add 'public' modifier to early Task classes

[hsuryawirawan] Rename Tests classes to TaskTest

[hsuryawirawan] Add branching Java kata

[hsuryawirawan] Add branching Python kata

[hsuryawirawan] Add TextIO Read Java kata

[hsuryawirawan] Add the Kata descriptions for Branching and TextIO Read

[hsuryawirawan] Add TextIO ReadFromText Python kata

[hsuryawirawan] Add Kata description from Branching Python kata

[hsuryawirawan] Add Built-in IOs Java kata

[hsuryawirawan] Add Built-in IOs Python kata

[lukasz.gajowy] [BEAM-7368] Bugfix: create proper amount of VMs for Flink Dataproc

[lukasz.gajowy] [BEAM-7368] Remove deps to external spark config in Flink init action

[hsuryawirawan] Add missing copyright for Python kata "Built-in IOs" test file

[hsuryawirawan] Add rat exclusion for Katas IO txt files

[hsuryawirawan] Removed unused entry in Java kata study_project.xml

[juta.staes] [BEAM-6769] allow read None bytes

[je.ik] [BEAM-7529] Add Sums.ofFloats() and Sums.ofDoubles()

[neville] [BEAM-7533] Fix CoderRegistry for Float

[ehudm] [BEAM-7531] Remove and disallow KMS key setting

[ehudm] Reduce gradle build verbosity

[mxm] [website] Add artifact staging document to design documents

[aromanenko.dev] [BEAM-7445] Add search field to Beam website

[github] [BEAM-7510] Fixing fileio tests checking JSON serialization (flaky)

[github] Remove Python batch support for Pub/Sub from docs

[kcweaver] [BEAM-7221] portable Spark: timers

[angoenka] [BEAM-7407] Adds portable wordcount test for Python 3 with Flink runner.

[rezarokni] Fixed spaces for author in looping blog. Added twitter handle for

[github] Use code.jquery.com for jQuery CDN

[kanterov] Merge pull request #8552: [BEAM-7268] make sorter extension Hadoop-free

[github] Add new proposal to design docs

[github] Documenting and refactoring fn api runner code (#8659)

[valentyn] Read YAML as binary stream in standard_coders_test.

[lcwik] [BEAM-7541] provide runner & filesystem deps for IOIT in runtime

[github] [BEAM-5315] Fix Py3 incompatibility in tfrecordio_test.

[github] Adding helper functions, and documentation to fn_api_runner classes.

[melissapa] [BEAM-7033] Updates for reading with query using BigQuery Storage API

[pekopeko] address github comment

[kcweaver] [BEAM-7553] add portability to Spark webpage

[iemejia] [BEAM-7043] Add DynamoDBIO

[thw] [BEAM-7126] Fix StateRequestHandler type variables to allow for proper

[kamil.wasilewski] [BEAM-7402] Added a performance test for BigQuery IO read

[mxm] [BEAM-7551] Checkpoint Flink's ImpulseSourceFunction

[david.moravek] [BEAM-7561] HdfsFileSystems.match can match a directory.

[kanterov] Merge pull request #8858: [BEAM-7542] Fix faulty cast in BigQueryUtils

[robertwb] Remove unused **kwargs arguments for various transforms.

[lukasz.gajowy] Revert "Merge pull request #8561: [BEAM-6627] Add item and byte counters

[lukasz.gajowy] Revert "Merge pull request #8400: [BEAM-6627] Added byte and item

[hsuryawirawan] Add README files on how to setup the project for both Java and Python

[github] Adding a Link to The Doc for Cost Estimation

[ehudm] [BEAM-6877] Fix trivial_inference for Python 3.x

[millsd] Use StateTags.ID_EQUIVALENCE when using comparing StateTags

[ehudm] Address review comments

[github] Add link to type hints design doc

[iemejia] [BEAM-7450] Support unbounded reads with HCatalogIO

[valentyn] Add Python 3.6, 3.7 to the list of supported version classifiers.

[chamikara] Fixes filesystem_test for Windows.

[github] use https

[mxm] [BEAM-7144] Fix for rescaling problem on Flink >= 1.6

[je.ik] [BEAM-7543] ReduceByKey.combineBy must accept BinaryFunction<V, V, V>

[alireza4263] [BEAM-7513] Implements row estimation for BigQuery.

[aaltay] [BEAM-6777] Let HealthzServlet respond actual health information of SDK

[amaliujia] support Datetime value conversion.

[melissapa] Merge pull request #8836: Add Beam Katas to the website's "Learning

[amaliujia] [BEAM-7461] disalbe flaky tests due to flaky

[github] Revert "[BEAM-7513] Adding Row Count for Bigquery Table"

[ehudm] [BEAM-7579] Use bucket with default key in ITs

[github] [BEAM-7467] Add dependency classifier to published pom (#8868)

[github] Fixing file naming for windows (#8870)

[zyichi] [BEAM-5148] Add MongoDB IO to python sdk

[ehudm] Allow more time for IT pipeline to run

[ehudm] Address review comments

[sniemitz] [BEAM-7507] Avoid more String.format and ByteString.toStringUtf8 in hot

[relax] [BEAM-6674] Add schema support to JdbcIO read (#8725)

[david.moravek] [BEAM-7561] Validate correct directory matching in LocalFileSystemsTest

[owenzhang1990] [BEAM-7467] Do not run gearpump pipeline remotely by default

[je.ik] [BEAM-7543] deprecate old #combineBy(Sums.ofLongs(),

[zyichi] Fix erros in py sdk io utils and add unit tests

[dcavazos] Add common code infrastructure for element-wise snippets

[alireza4263] [BEAM-7513] Adding RowCount to BigQueryTable.

[logan.hauspie.pro] [BEAM-5806] Update PubsubIO to be able to change the PubsubClientFactory

[aaltay] [BEAM-6955] Use base version component of Beam Python SDK version when

[kedin] Moving to 2.15.0-SNAPSHOT on master branch.

[dcavazos] Add Python snippet for Filter transform

[dcavazos] Add Python snippet for Map transform

[dcavazos] Add Python snippet for Keys, Values, KvSwap, and ToString transform

[github] [BEAM-7603] Adding support for ValueProvider in Custom GCS Temp Location

[github] Add link to gRPC.io on website portability page

[agoos] [BEAM7597] Fix Typo on website

[github] [BEAM-4046, BEAM-7307] revert PR 8581, migrate jenkins tests to use

[lukecwik] Make PubsubIO#withCoderAndParseFn public (#8879)

[ttanay100] Add to docstring of advance_watermark_to_infinity

[mxm] Fixed typo and removed whitespace

[github] Add a timeout to urlopen calls

[dcavazos] Add Python snippet for FlatMap transform

[aaltay] [Beam-6696] GroupIntoBatches transform for Python SDK (#8914)

[gunnar.schulze] [BEAM-7572] ApproximateUnique.Globally and ApproximateUnique.PerKey

[lukasz.gajowy] [BEAM-7307] Fix load tests failures due to mistakes in PR 8881

[aryan.naraghi] Implement splitAtFraction for the BigQuery Storage API

[github] Drop experimental from python streaming in the doc

[markliu] [BEAM-7598] Do not build Python tar file in run_integration_test.sh

[dcavazos] Add Python snippet for ParDo transform

[aaltay] BEAM-7141: add key value timer callback (#8739)

[iemejia] [BEAM-7606] Fix JDBC time conversion tests

[robertwb] Update portable schema representation and java SchemaTranslation (#8853)

[juta.staes] [BEAM-7326] add documentation bigquery data types

[juta.staes] [BEAM-5315] improve test coverage bigquery special chars

[lcwik] [BEAM-4948, BEAM-6267, BEAM-5559, BEAM-7289] Update the version of guava

[aaltay] [BEAM-7475] update wordcount example (#8803)

[github] [BEAM-7013] Add HLL doc link to Beam website

[heejong] [BEAM-7424] Retry HTTP 429 errors from GCS

[markliu] [BEAM-4046, BEAM-7527] Fix benchmark with correct Gradle project

[zyichi] [BEAM-7586] Add Integration test for python mongodb io

[valentyn] Match Python 3 warning message in __init__.py with the one in setup.py.

[kedin] Spotless config update to include java files only under src directory

[chamikara] [BEAM-7548] fix flaky tests for ApproximateUnique (#8948)

[samuelw] [BEAM-7547] Avoid WindmillStateCache cache hits for stale work.

[iemejia] [BEAM-7640] Create amazon-web-services2 module and AwsOptions

[alireza4263] [BEAM-7545] Adding RowCount to TextTable.

[kcweaver] [BEAM-6692] portable Spark: reshuffle translation

[chamikara] [BEAM-7548] Fix flaky tests for ApproximateUnique (#8960)

[iemejia] [BEAM-7589] Use only one KinesisProducer instance per JVM

[iemejia] [BEAM-7589] Make KinesisIOIT compatible with all other ITs

[daniel.o.programmer] Update python containers to beam-master-20190605

[hannahjiang] BEAM-3645 add ParallelBundleProcessor

[hannahjiang] BEAM-3645 reflect comments

[hannahjiang] BEAM-3645 add changes from review comments

[hannahjiang] BEAM-3645 add thread lock when generating process_bundle_id

[github] Tiny typo fix

[kamil.wasilewski] [BEAM-7536] Fixed BQ dataset name in collecting Load Tests metrics

[kamil.wasilewski] [BEAM-7504] Added top_count parameter

[lukasz.gajowy] [BEAM-4420] Allow connecting to zookeeper using external ip

[kamil.wasilewski] [BEAM-7504] Create Combine Python Load Test Jenkins job

[iemejia] [BEAM-7640] Rename the package name for amazon-web-services2 from aws to

[lukasz.gajowy] [BEAM-4420] Add KafkaIO integration test pipeline

------------------------------------------
[...truncated 146.34 KB...]
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {
          "side0-Combine with Top 0/InjectDefault": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "SideInput-s8"
          }
        }, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "Combine with Top 0/InjectDefault.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s7"
        }, 
        "serialized_fn": "<string of 1824 bytes>", 
        "user_name": "Combine with Top 0/InjectDefault/InjectDefault"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s10", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "_GetElement", 
            "type": "STRING", 
            "value": "apache_beam.testing.load_tests.combine_test._GetElement"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "Consume 0.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s9"
        }, 
        "serialized_fn": "<string of 432 bytes>", 
        "user_name": "Consume 0"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s11", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "MeasureTime", 
            "type": "STRING", 
            "value": "apache_beam.testing.load_tests.load_test_metrics_utils.MeasureTime"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "Measure time: End 0.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s10"
        }, 
        "serialized_fn": "<string of 504 bytes>", 
        "user_name": "Measure time: End 0"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2019-07-03T15:50:41.015150Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2019-07-03_08_50_39-5054764375871059839'
 location: u'us-central1'
 name: u'load-tests-python-dataflow-batch-combine-3-0703150146'
 projectId: u'apache-beam-testing'
 stageStates: []
 startTime: u'2019-07-03T15:50:41.015150Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-07-03_08_50_39-5054764375871059839]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-07-03_08_50_39-5054764375871059839?project=apache-beam-testing
root: INFO: Job 2019-07-03_08_50_39-5054764375871059839 is in state JOB_STATE_PENDING
root: INFO: 2019-07-03T15:50:39.993Z: JOB_MESSAGE_WARNING: The requested max number of workers (5) is ignored as autoscaling is explicitly disabled (autoscalingAlgorithm=NONE).
root: INFO: 2019-07-03T15:50:43.643Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-07-03T15:50:44.130Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-a.
root: INFO: 2019-07-03T15:50:44.794Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-07-03T15:50:44.836Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
root: INFO: 2019-07-03T15:50:44.879Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-07-03T15:50:44.979Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-07-03T15:50:45.044Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-07-03T15:50:45.090Z: JOB_MESSAGE_DETAILED: Fusing consumer Measure time: Start into Read
root: INFO: 2019-07-03T15:50:45.129Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/GroupByKey+Combine with Top 0/CombinePerKey/Combine/Partial into Combine with Top 0/KeyWithVoid
root: INFO: 2019-07-03T15:50:45.182Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/Combine/Extract into Combine with Top 0/CombinePerKey/Combine
root: INFO: 2019-07-03T15:50:45.241Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/Combine into Combine with Top 0/CombinePerKey/GroupByKey/Read
root: INFO: 2019-07-03T15:50:45.282Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/KeyWithVoid into Measure time: Start
root: INFO: 2019-07-03T15:50:45.333Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/GroupByKey/Write into Combine with Top 0/CombinePerKey/GroupByKey/Reify
root: INFO: 2019-07-03T15:50:45.378Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/CombinePerKey/GroupByKey/Reify into Combine with Top 0/CombinePerKey/GroupByKey+Combine with Top 0/CombinePerKey/Combine/Partial
root: INFO: 2019-07-03T15:50:45.425Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/UnKey into Combine with Top 0/CombinePerKey/Combine/Extract
root: INFO: 2019-07-03T15:50:45.477Z: JOB_MESSAGE_DETAILED: Fusing consumer Measure time: End 0 into Consume 0
root: INFO: 2019-07-03T15:50:45.533Z: JOB_MESSAGE_DETAILED: Fusing consumer Consume 0 into Combine with Top 0/InjectDefault/InjectDefault
root: INFO: 2019-07-03T15:50:45.581Z: JOB_MESSAGE_DETAILED: Fusing consumer Combine with Top 0/InjectDefault/InjectDefault into Combine with Top 0/DoOnce/Read
root: INFO: 2019-07-03T15:50:45.623Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
root: INFO: 2019-07-03T15:50:45.674Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-07-03T15:50:45.716Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-07-03T15:50:45.757Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-07-03T15:50:45.983Z: JOB_MESSAGE_DEBUG: Executing wait step start23
root: INFO: Job 2019-07-03_08_50_39-5054764375871059839 is in state JOB_STATE_RUNNING
root: INFO: 2019-07-03T15:50:46.111Z: JOB_MESSAGE_BASIC: Executing operation Combine with Top 0/CombinePerKey/GroupByKey/Create
root: INFO: 2019-07-03T15:50:46.183Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-07-03T15:50:46.233Z: JOB_MESSAGE_BASIC: Starting 5 workers in us-central1-a...
root: INFO: 2019-07-03T15:50:46.303Z: JOB_MESSAGE_BASIC: Finished operation Combine with Top 0/CombinePerKey/GroupByKey/Create
root: INFO: 2019-07-03T15:50:46.414Z: JOB_MESSAGE_DEBUG: Value "Combine with Top 0/CombinePerKey/GroupByKey/Session" materialized.
root: INFO: 2019-07-03T15:50:46.525Z: JOB_MESSAGE_BASIC: Executing operation Read+Measure time: Start+Combine with Top 0/KeyWithVoid+Combine with Top 0/CombinePerKey/GroupByKey+Combine with Top 0/CombinePerKey/Combine/Partial+Combine with Top 0/CombinePerKey/GroupByKey/Reify+Combine with Top 0/CombinePerKey/GroupByKey/Write
root: INFO: 2019-07-03T15:51:49.574Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 4 based on the rate of progress in the currently running step(s).
root: INFO: 2019-07-03T15:51:49.612Z: JOB_MESSAGE_DETAILED: Resized worker pool to 4, though goal was 5.  This could be a quota issue.
root: INFO: 2019-07-03T15:51:54.897Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 5 based on the rate of progress in the currently running step(s).
root: INFO: 2019-07-03T15:52:06.737Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
root: INFO: 2019-07-03T15:52:06.814Z: JOB_MESSAGE_DETAILED: Workers have started successfully.
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_LoadTests_Python_Combine_Dataflow_Batch/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 137.115s

FAILED (errors=1)

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_Combine_Dataflow_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 49

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 2m 27s
4 actionable tasks: 3 executed, 1 up-to-date

Publishing build scan...
https://gradle.com/s/n7obhoszqlrsw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org