You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/01/29 00:33:47 UTC
Build failed in Jenkins: beam_PostCommit_Python_Verify #7224
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/7224/display/redirect?page=changes>
Changes:
[gleb] [BEAM-6497] Upgrade testcontainers
[amaliujia] [BEAM-6513] Fix NPE in BeamEnumerableConveter.
------------------------------------------
[...truncated 430.76 KB...]
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "To Entity.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s2"
},
"serialized_fn": "<string of 1464 bytes>",
"user_name": "To Entity"
}
},
{
"kind": "ParallelDo",
"name": "s4",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "to_upsert_mutation"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "Write to Datastore/Convert to Mutation.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s3"
},
"serialized_fn": "<string of 1244 bytes>",
"user_name": "Write to Datastore/Convert to Mutation"
}
},
{
"kind": "ParallelDo",
"name": "s5",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "DatastoreWriteFn",
"type": "STRING",
"value": "apache_beam.io.gcp.datastore.v1.datastoreio.DatastoreWriteFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "Write to Datastore/Write Mutation to Datastore.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s4"
},
"serialized_fn": "<string of 6532 bytes>",
"user_name": "Write to Datastore/Write Mutation to Datastore"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
createTime: u'2019-01-29T00:15:59.938092Z'
currentStateTime: u'1970-01-01T00:00:00Z'
id: u'2019-01-28_16_15_58-11535588838532301848'
location: u'us-central1'
name: u'beamapp-jenkins-0129001553-564053'
projectId: u'apache-beam-testing'
stageStates: []
startTime: u'2019-01-29T00:15:59.938092Z'
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-01-28_16_15_58-11535588838532301848]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_15_58-11535588838532301848?project=apache-beam-testing
root: INFO: Job 2019-01-28_16_15_58-11535588838532301848 is in state JOB_STATE_RUNNING
root: INFO: 2019-01-29T00:15:58.904Z: JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 2019-01-28_16_15_58-11535588838532301848. The number of workers will be between 1 and 1000.
root: INFO: 2019-01-29T00:15:58.951Z: JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 2019-01-28_16_15_58-11535588838532301848.
root: INFO: 2019-01-29T00:16:03.931Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-01-29T00:16:06.076Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-f.
root: INFO: 2019-01-29T00:16:06.834Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-01-29T00:16:06.883Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
root: INFO: 2019-01-29T00:16:06.937Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-01-29T00:16:06.982Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-01-29T00:16:07.032Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-01-29T00:16:07.083Z: JOB_MESSAGE_DETAILED: Fusing consumer To String into Input/Read
root: INFO: 2019-01-29T00:16:07.126Z: JOB_MESSAGE_DETAILED: Fusing consumer Write to Datastore/Convert to Mutation into To Entity
root: INFO: 2019-01-29T00:16:07.169Z: JOB_MESSAGE_DETAILED: Fusing consumer To Entity into To String
root: INFO: 2019-01-29T00:16:07.212Z: JOB_MESSAGE_DETAILED: Fusing consumer Write to Datastore/Write Mutation to Datastore into Write to Datastore/Convert to Mutation
root: INFO: 2019-01-29T00:16:07.263Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
root: INFO: 2019-01-29T00:16:07.307Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-01-29T00:16:07.350Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-01-29T00:16:07.372Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-01-29T00:16:07.560Z: JOB_MESSAGE_DEBUG: Executing wait step start3
root: INFO: 2019-01-29T00:16:07.656Z: JOB_MESSAGE_BASIC: Executing operation Input/Read+To String+To Entity+Write to Datastore/Convert to Mutation+Write to Datastore/Write Mutation to Datastore
root: INFO: 2019-01-29T00:16:07.712Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-01-29T00:16:07.756Z: JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-f...
root: INFO: 2019-01-29T00:16:35.055Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 0 based on the rate of progress in the currently running step(s).
root: INFO: 2019-01-29T00:20:46.905Z: JOB_MESSAGE_ERROR: Startup of the worker pool in zone us-central1-f failed to bring up any of the desired 1 workers. ZONE_RESOURCE_POOL_EXHAUSTED_WITH_DETAILS: The zone 'projects/apache-beam-testing/zones/us-central1-f' does not have enough resources available to fulfill the request. '(resource type:pd-standard)'.
root: INFO: 2019-01-29T00:20:46.950Z: JOB_MESSAGE_ERROR: Workflow failed.
root: INFO: 2019-01-29T00:20:47.129Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2019-01-29T00:20:47.248Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown.
root: INFO: 2019-01-29T00:20:47.294Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2019-01-29T00:20:57.739Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2019-01-29T00:20:57.813Z: JOB_MESSAGE_DEBUG: Tearing down pending resources...
root: INFO: Job 2019-01-28_16_15_58-11535588838532301848 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 25 tests in 1963.491s
FAILED (errors=2)
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_11-11727449319289504858?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_15_49-6076934208658781741?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_10-12171153624806607855?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_10-14655584162582924270?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_11-17557616721073621738?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_13_34-16472480060947822327?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_19_58-8568388869577934712?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_27_32-11046474615481520434?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_10-18040888023890492425?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_06_44-12947325951682453911?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_16_26-14256921523204636930?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_24_07-8297770678264826224?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_09-4776643139496807942?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_08_00-8397597758696897334?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_15_58-11535588838532301848?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_09-4175208168581871724?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_09_12-2945928238047091748?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_16_44-4710928883396691880?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_01_10-13802233784993623460?project=apache-beam-testing.
Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-01-28_16_08_37-15873422496318915146?project=apache-beam-testing.
> Task :beam-sdks-python:postCommitIT FAILED
FAILURE: Build failed with an exception.
* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/src/sdks/python/build.gradle'> line: 278
* What went wrong:
Execution failed for task ':beam-sdks-python:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
BUILD FAILED in 36m 50s
6 actionable tasks: 6 executed
Publishing build scan...
https://gradle.com/s/wvcbtwzzv72zm
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Jenkins build is back to normal : beam_PostCommit_Python_Verify
#7225
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/7225/display/redirect>
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org